
Advertising on the Telegram channel «Artificial Intelligence»
🔰 Machine Learning & Artificial Intelligence Free Resources
🔰 Advanced Data Science and Deep Learning Concepts
🔰 Build Chatbots & Large Language Models
🔰 Learn PyTorch & Tensorflow
Channel statistics
A beginner-friendly 21-lesson course by Microsoft that teaches how to build real generative AI apps—from prompts to RAG, agents, and deployment.2️⃣ rasbt/LLMs-from-scratch
Learn how LLMs actually work by building a GPT-style model step by step in pure PyTorch—ideal for deeply understanding LLM internals.3️⃣ DataTalksClub/llm-zoomcamp
A free 10-week, hands-on course focused on production-ready LLM applications, especially RAG systems built over your own data.4️⃣ Shubhamsaboo/awesome-llm-apps
A curated collection of real, runnable LLM applications showcasing agents, RAG pipelines, voice AI, and modern agentic patterns.5️⃣ panaversity/learn-agentic-ai
A practical program for designing and scaling cloud-native, production-grade agentic AI systems using Kubernetes, Dapr, and multi-agent workflows.6️⃣ dair-ai/Mathematics-for-ML
A carefully curated library of books, lectures, and papers to master the mathematical foundations behind machine learning and deep learning.7️⃣ ashishpatel26/500-AI-ML-DL-Projects-with-code
A massive collection of 500+ AI project ideas with code across computer vision, NLP, healthcare, recommender systems, and real-world ML use cases.8️⃣ armankhondker/awesome-ai-ml-resources
A clear 2025 roadmap that guides learners from beginner to advanced AI with curated resources and career-focused direction.9️⃣ spmallick/learnopencv
One of the best hands-on repositories for computer vision, covering OpenCV, YOLO, diffusion models, robotics, and edge AI.🔟 x1xhlol/system-prompts-and-models-of-ai-tools
A deep dive into how real AI tools are built, featuring 30K+ lines of system prompts, agent designs, and production-level AI patterns.
import tensorflow as tf
from tensorflow.keras import layers, models
import matplotlib.pyplot as plt
{}
Step 2. Load and Prepare Data
(x_train, y_train), (x_test, y_test) = tf.keras.datasets.mnist.load_data()
x_train = x_train / 255.0
x_test = x_test / 255.0
x_train = x_train.reshape(-1, 28, 28, 1)
x_test = x_test.reshape(-1, 28, 28, 1)
{}
Step 3. Build CNN Model
model = models.Sequential([
layers.Conv2D(32, (3,3), activation="relu", input_shape=(28,28,1)),
layers.MaxPooling2D((2,2)),
layers.Conv2D(64, (3,3), activation="relu"),
layers.MaxPooling2D((2,2)),
layers.Flatten(),
layers.Dense(128, activation="relu"),
layers.Dense(10, activation="softmax")
])
{}
Step 4. Compile Model
model.compile(
optimizer="adam",
loss="sparse_categorical_crossentropy",
metrics=["accuracy"]
)
{}
Step 5. Train Model
model.fit(
x_train, y_train,
epochs=5,
validation_split=0.1
)
{}
Step 6. Evaluate Model
test_loss, test_accuracy = model.evaluate(x_test, y_test)
print("Test accuracy:", test_accuracy)
{}
Expected output
Test accuracy around 0.98
Stable validation curve
Fast training on CPU or GPU
Testing with Custom Image
Convert image to grayscale
Resize to 28 × 28
Normalize pixel values
Pass through model.predict
Common Mistakes
Skipping normalization
Wrong image shape
Using RGB instead of grayscale
Portfolio Value
- Shows computer vision basics
- Demonstrates CNN understanding
- Easy to explain in interviews
- Strong beginner-to-intermediate project
Double Tap ♥️ For Part-3f(x) = max(0, x)
✔️ Fast
✔️ Prevents vanishing gradients
❌ Can "die" (output 0 for all inputs if weights go bad)
b) Sigmoid
f(x) = 1 / (1 + exp(-x))
✔️ Good for binary output
❌ Causes vanishing gradient
❌ Not zero-centered
c) Tanh (Hyperbolic Tangent)
f(x) = (exp(x) - exp(-x)) / (exp(x) + exp(-x))
✔️ Outputs between -1 and 1
✔️ Zero-centered
❌ Still suffers vanishing gradient
d) Leaky ReLU
f(x) = x if x > 0 else 0.01 * x
✔️ Fixes dying ReLU issue
✔️ Allows small gradient for negative inputs
e) Softmax
Used in final layer for multi-class classification
✔️ Converts outputs into probability distribution
✔️ Sum of outputs = 1
3️⃣ Where to Use What?
• ReLU → Hidden layers (default choice)
• Sigmoid → Output layer for binary classification
• Tanh → Hidden layers (sometimes better than sigmoid)
• Softmax → Final layer for multi-class problems
🧪 Try This:
Build a model with:
• ReLU in hidden layers
• Softmax in output
• Use it for classifying handwritten digits (MNIST)
💬 Tap ❤️ for more!Reviews channel
7 total reviews
- Added: Newest first
- Added: Oldest first
- Rating: High to low
- Rating: Low to high
Catalog of Telegram Channels for Native Placements
Advertising on the Telegram channel «Artificial Intelligence» is a Telegram channel in the category «Образование», offering effective formats for placing advertising posts on TG. The channel has 49.9K subscribers and provides quality content. The advertising posts on the channel help brands attract audience attention and increase reach. The channel's rating is 13.1, with 7 reviews and an average score of 5.0.
You can launch an advertising campaign through the Telega.in service, choosing a convenient format for placement. The Platform provides transparent cooperation conditions and offers detailed analytics. The placement cost is 18.0 ₽, and with 20 completed requests, the channel has established itself as a reliable partner for advertising on Telegram. Place integrations today and attract new clients!
You will be able to add channels from the catalog to the cart again.
Комментарий