• 🌙 Community Spirit

    Ramadan Mubarak! To honor this month, Crax has paused NSFW categories. Wishing you peace and growth!

IT & Software Other AI Engineer (1 Viewer)

Currently reading:
 IT & Software Other AI Engineer (1 Viewer)

Covers web development, programming, AI, cloud computing, DevOps, and cybersecurity.
Recently searched:

kashifpcshop

Member
LV
3
Joined
Sep 23, 2021
Threads
60
Likes
36
Awards
8
Credits
3,649©
Cash
0$

Learn by Doing.​

Become an AI Engineer.


6 Weeks · Cohort-based Course, Next cohort Nov 8—Dec 14, 2025
Course page: bytebyteai

1762696528176

1762696539304

1762696550606

1762696562094

1762696574035

1762696584693



Taught by Best-Selling

best-selling-underline.BDIoJigS_2s5Ai.svg

Author Ali Aminian​

1762696611559

Meet Your Instructor​

Ali Aminian​

1762696641364

Ali Aminian is a best-selling author of multiple books on machine learning and generative AI. With over a decade of experience at leading tech companies, he has built AI systems that are intelligent, safe, and efficient. He also contributes to AI courses at Stanford University, combining technical expertise with a passion for teaching.
Adobe
Google
Stanford


Course Outline (Project-Based Learning)​


Project 1

Build an LLM Playground​

1762696690676

LLM Overview and Foundations
Pre-Training
  • Data collection (manual crawling, Common Crawl)
  • Data cleaning (RefinedWeb, Dolma, FineWeb)
  • Tokenization (e.g., BPE)
  • Architecture (neural networks, Transformers, GPT family, Llama family)
  • Text generation (greedy and beam search, top-k, top-p)
Post-Training
  • SFT
  • RL and RLHF (verifiable tasks, reward models, PPO, etc.)
Evaluation
  • Traditional metrics
  • Task-specific benchmarks
  • Human evaluation and leaderboards
Chatbots' Overall Design

Project 2

Build a Customer Support Chatbot using RAGs and Prompt Engineering​

1762696707809

Overview of Adaptation Techniques
Finetuning
  • Parameter-efficient fine-tuning (PEFT)
  • Adapters and LoRA
Prompt Engineering
  • Few-shot and zero-shot prompting
  • Chain-of-thought prompting
  • Role-specific and user-context prompting
RAGs Overview
Retrieval
  • Document parsing (rule-based, AI-based) and chunking strategies
  • Indexing (keyword, full-text, knowledge-based, vector-based, embedding models)
Generation
  • Search methods (exact and approximate nearest neighbor)
  • Prompt engineering for RAGs
RAFT: Training technique for RAGs
Evaluation (context relevance, faithfulness, answer correctness)
RAGs' Overall Design

Project 3

Build an "Ask-the-Web" Agent similar to Perplexity with Tool calling​

1762696725996

Agents Overview
  • Agents vs. agentic systems vs. LLMs
  • Agency levels (e.g., workflows, multi-step agents)
Workflows
  • Prompt chaining
  • Routing
  • Parallelization (sectioning, voting)
  • Reflection
  • Orchestration-worker
Tools
  • Tool calling
  • Tool formatting
  • Tool execution
  • MCP
Multi-Step Agents
  • Planning autonomy
  • ReACT
  • Reflexion, ReWOO, etc.
  • Tree search for agents
Multi-Agent Systems (challenges, use-cases, A2A protocol)
Evaluation of agents

Project 4

Build "Deep Research" Capability with Web Search and Reasoning Models​

1762696740588

Reasoning and Thinking LLMs
  • Overview of reasoning models like OpenAI's "o" family and DeepSeek-R1
Inference-time Techniques
  • Inferece-time scaling
  • CoT prompting
  • Self-consistency
  • Sequential revision
  • Tree of Thoughts (ToT)
  • Search against a verifier
Training-time techniques
  • SFT on reasoning data (e.g., STaR)
  • Reinforcement learning with a verifier
  • Reward modeling (ORM, PRM)
  • Self-refinement
  • Internalizing search (e.g., Meta-CoT)

Project 5

Build a Multi-modal Generation Agent​

1762696754932

Overview of Image and Video Generation
  • VAE
  • GANs
  • Auto-regressive models
  • Diffusion models
Text-to-Image (T2I)
  • Data preparation
  • Diffusion architectures (U-Net, DiT)
  • Diffusion training (forward process, backward process)
  • Diffusion sampling
  • Evaluation (image quality, diversity, image-text alignment, IS, FID, and CLIP score)
Text-to-Video (T2V)
  • Latent-diffusion modeling (LDM) and compression networks
  • Data preparation (filtering, standardization, video latent caching)
  • DiT architecture for videos
  • Large-scale training challenges
  • T2V's overall system

Project 6

Capstone Project​

1762696770536

  • Choose your own idea
  • Build with techniques from the course
  • Get real-time feedback from the instructor as you build
  • Demo + feedback session


Link:

1762696668447
 
Last edited:

Xcosine

Member
LV
1
Joined
Jun 6, 2024
Threads
11
Likes
25
Awards
5
Credits
2,730©
Cash
0$

Learn by Doing.​

Become an AI Engineer.


6 Weeks · Cohort-based Course, Next cohort Nov 8—Dec 14, 2025
Course page: bytebyteai

View attachment 295755
View attachment 295756
View attachment 295757
View attachment 295758
View attachment 295759
View attachment 295760


Taught by Best-Selling

best-selling-underline.BDIoJigS_2s5Ai.svg

Author Ali Aminian​

Meet Your Instructor​

Ali Aminian​

View attachment 295762
Ali Aminian is a best-selling author of multiple books on machine learning and generative AI. With over a decade of experience at leading tech companies, he has built AI systems that are intelligent, safe, and efficient. He also contributes to AI courses at Stanford University, combining technical expertise with a passion for teaching.
Adobe
Google
Stanford


Course Outline (Project-Based Learning)​


Project 1

Build an LLM Playground​

View attachment 295764
LLM Overview and Foundations
Pre-Training
  • Data collection (manual crawling, Common Crawl)
  • Data cleaning (RefinedWeb, Dolma, FineWeb)
  • Tokenization (e.g., BPE)
  • Architecture (neural networks, Transformers, GPT family, Llama family)
  • Text generation (greedy and beam search, top-k, top-p)
Post-Training
  • SFT
  • RL and RLHF (verifiable tasks, reward models, PPO, etc.)
Evaluation
  • Traditional metrics
  • Task-specific benchmarks
  • Human evaluation and leaderboards
Chatbots' Overall Design

Project 2

Build a Customer Support Chatbot using RAGs and Prompt Engineering​

View attachment 295765
Overview of Adaptation Techniques
Finetuning
  • Parameter-efficient fine-tuning (PEFT)
  • Adapters and LoRA
Prompt Engineering
  • Few-shot and zero-shot prompting
  • Chain-of-thought prompting
  • Role-specific and user-context prompting
RAGs Overview
Retrieval
  • Document parsing (rule-based, AI-based) and chunking strategies
  • Indexing (keyword, full-text, knowledge-based, vector-based, embedding models)
Generation
  • Search methods (exact and approximate nearest neighbor)
  • Prompt engineering for RAGs
RAFT: Training technique for RAGs
Evaluation (context relevance, faithfulness, answer correctness)
RAGs' Overall Design

Project 3

Build an "Ask-the-Web" Agent similar to Perplexity with Tool calling​

View attachment 295766
Agents Overview
  • Agents vs. agentic systems vs. LLMs
  • Agency levels (e.g., workflows, multi-step agents)
Workflows
  • Prompt chaining
  • Routing
  • Parallelization (sectioning, voting)
  • Reflection
  • Orchestration-worker
Tools
  • Tool calling
  • Tool formatting
  • Tool execution
  • MCP
Multi-Step Agents
  • Planning autonomy
  • ReACT
  • Reflexion, ReWOO, etc.
  • Tree search for agents
Multi-Agent Systems (challenges, use-cases, A2A protocol)
Evaluation of agents

Project 4

Build "Deep Research" Capability with Web Search and Reasoning Models​

View attachment 295767
Reasoning and Thinking LLMs
  • Overview of reasoning models like OpenAI's "o" family and DeepSeek-R1
Inference-time Techniques
  • Inferece-time scaling
  • CoT prompting
  • Self-consistency
  • Sequential revision
  • Tree of Thoughts (ToT)
  • Search against a verifier
Training-time techniques
  • SFT on reasoning data (e.g., STaR)
  • Reinforcement learning with a verifier
  • Reward modeling (ORM, PRM)
  • Self-refinement
  • Internalizing search (e.g., Meta-CoT)

Project 5

Build a Multi-modal Generation Agent​

View attachment 295768
Overview of Image and Video Generation
  • VAE
  • GANs
  • Auto-regressive models
  • Diffusion models
Text-to-Image (T2I)
  • Data preparation
  • Diffusion architectures (U-Net, DiT)
  • Diffusion training (forward process, backward process)
  • Diffusion sampling
  • Evaluation (image quality, diversity, image-text alignment, IS, FID, and CLIP score)
Text-to-Video (T2V)
  • Latent-diffusion modeling (LDM) and compression networks
  • Data preparation (filtering, standardization, video latent caching)
  • DiT architecture for videos
  • Large-scale training challenges
  • T2V's overall system

Project 6

Capstone Project​

View attachment 295769
  • Choose your own idea
  • Build with techniques from the course
  • Get real-time feedback from the instructor as you build
  • Demo + feedback session


Link:
* Hidden text: cannot be quoted. *

View attachment 295763
THanks
 

sismuk903

Member
LV
2
Joined
Jan 8, 2023
Threads
4
Likes
1
Awards
6
Credits
8,402©
Cash
0$

Learn by Doing.​

Become an AI Engineer.


6 Weeks · Cohort-based Course, Next cohort Nov 8—Dec 14, 2025
Course page: bytebyteai

View attachment 295755
View attachment 295756
View attachment 295757
View attachment 295758
View attachment 295759
View attachment 295760


Taught by Best-Selling

best-selling-underline.BDIoJigS_2s5Ai.svg

Author Ali Aminian​

Meet Your Instructor​

Ali Aminian​

View attachment 295762
Ali Aminian is a best-selling author of multiple books on machine learning and generative AI. With over a decade of experience at leading tech companies, he has built AI systems that are intelligent, safe, and efficient. He also contributes to AI courses at Stanford University, combining technical expertise with a passion for teaching.
Adobe
Google
Stanford


Course Outline (Project-Based Learning)​


Project 1

Build an LLM Playground​

View attachment 295764
LLM Overview and Foundations
Pre-Training
  • Data collection (manual crawling, Common Crawl)
  • Data cleaning (RefinedWeb, Dolma, FineWeb)
  • Tokenization (e.g., BPE)
  • Architecture (neural networks, Transformers, GPT family, Llama family)
  • Text generation (greedy and beam search, top-k, top-p)
Post-Training
  • SFT
  • RL and RLHF (verifiable tasks, reward models, PPO, etc.)
Evaluation
  • Traditional metrics
  • Task-specific benchmarks
  • Human evaluation and leaderboards
Chatbots' Overall Design

Project 2

Build a Customer Support Chatbot using RAGs and Prompt Engineering​

View attachment 295765
Overview of Adaptation Techniques
Finetuning
  • Parameter-efficient fine-tuning (PEFT)
  • Adapters and LoRA
Prompt Engineering
  • Few-shot and zero-shot prompting
  • Chain-of-thought prompting
  • Role-specific and user-context prompting
RAGs Overview
Retrieval
  • Document parsing (rule-based, AI-based) and chunking strategies
  • Indexing (keyword, full-text, knowledge-based, vector-based, embedding models)
Generation
  • Search methods (exact and approximate nearest neighbor)
  • Prompt engineering for RAGs
RAFT: Training technique for RAGs
Evaluation (context relevance, faithfulness, answer correctness)
RAGs' Overall Design

Project 3

Build an "Ask-the-Web" Agent similar to Perplexity with Tool calling​

View attachment 295766
Agents Overview
  • Agents vs. agentic systems vs. LLMs
  • Agency levels (e.g., workflows, multi-step agents)
Workflows
  • Prompt chaining
  • Routing
  • Parallelization (sectioning, voting)
  • Reflection
  • Orchestration-worker
Tools
  • Tool calling
  • Tool formatting
  • Tool execution
  • MCP
Multi-Step Agents
  • Planning autonomy
  • ReACT
  • Reflexion, ReWOO, etc.
  • Tree search for agents
Multi-Agent Systems (challenges, use-cases, A2A protocol)
Evaluation of agents

Project 4

Build "Deep Research" Capability with Web Search and Reasoning Models​

View attachment 295767
Reasoning and Thinking LLMs
  • Overview of reasoning models like OpenAI's "o" family and DeepSeek-R1
Inference-time Techniques
  • Inferece-time scaling
  • CoT prompting
  • Self-consistency
  • Sequential revision
  • Tree of Thoughts (ToT)
  • Search against a verifier
Training-time techniques
  • SFT on reasoning data (e.g., STaR)
  • Reinforcement learning with a verifier
  • Reward modeling (ORM, PRM)
  • Self-refinement
  • Internalizing search (e.g., Meta-CoT)

Project 5

Build a Multi-modal Generation Agent​

View attachment 295768
Overview of Image and Video Generation
  • VAE
  • GANs
  • Auto-regressive models
  • Diffusion models
Text-to-Image (T2I)
  • Data preparation
  • Diffusion architectures (U-Net, DiT)
  • Diffusion training (forward process, backward process)
  • Diffusion sampling
  • Evaluation (image quality, diversity, image-text alignment, IS, FID, and CLIP score)
Text-to-Video (T2V)
  • Latent-diffusion modeling (LDM) and compression networks
  • Data preparation (filtering, standardization, video latent caching)
  • DiT architecture for videos
  • Large-scale training challenges
  • T2V's overall system

Project 6

Capstone Project​

View attachment 295769
  • Choose your own idea
  • Build with techniques from the course
  • Get real-time feedback from the instructor as you build
  • Demo + feedback session


Link:
* Hidden text: cannot be quoted. *

View attachment 295763
Thanks a lot bro.
 

Maxactuel

Member
LV
2
Joined
Jul 22, 2023
Threads
2
Likes
4
Awards
5
Credits
2,169©
Cash
0$

Learn by Doing.​

Become an AI Engineer.


6 Weeks · Cohort-based Course, Next cohort Nov 8—Dec 14, 2025
Course page: bytebyteai

View attachment 295755
View attachment 295756
View attachment 295757
View attachment 295758
View attachment 295759
View attachment 295760


Taught by Best-Selling

best-selling-underline.BDIoJigS_2s5Ai.svg

Author Ali Aminian​

Meet Your Instructor​

Ali Aminian​

View attachment 295762
Ali Aminian is a best-selling author of multiple books on machine learning and generative AI. With over a decade of experience at leading tech companies, he has built AI systems that are intelligent, safe, and efficient. He also contributes to AI courses at Stanford University, combining technical expertise with a passion for teaching.
Adobe
Google
Stanford


Course Outline (Project-Based Learning)​


Project 1

Build an LLM Playground​

View attachment 295764
LLM Overview and Foundations
Pre-Training
  • Data collection (manual crawling, Common Crawl)
  • Data cleaning (RefinedWeb, Dolma, FineWeb)
  • Tokenization (e.g., BPE)
  • Architecture (neural networks, Transformers, GPT family, Llama family)
  • Text generation (greedy and beam search, top-k, top-p)
Post-Training
  • SFT
  • RL and RLHF (verifiable tasks, reward models, PPO, etc.)
Evaluation
  • Traditional metrics
  • Task-specific benchmarks
  • Human evaluation and leaderboards
Chatbots' Overall Design

Project 2

Build a Customer Support Chatbot using RAGs and Prompt Engineering​

View attachment 295765
Overview of Adaptation Techniques
Finetuning
  • Parameter-efficient fine-tuning (PEFT)
  • Adapters and LoRA
Prompt Engineering
  • Few-shot and zero-shot prompting
  • Chain-of-thought prompting
  • Role-specific and user-context prompting
RAGs Overview
Retrieval
  • Document parsing (rule-based, AI-based) and chunking strategies
  • Indexing (keyword, full-text, knowledge-based, vector-based, embedding models)
Generation
  • Search methods (exact and approximate nearest neighbor)
  • Prompt engineering for RAGs
RAFT: Training technique for RAGs
Evaluation (context relevance, faithfulness, answer correctness)
RAGs' Overall Design

Project 3

Build an "Ask-the-Web" Agent similar to Perplexity with Tool calling​

View attachment 295766
Agents Overview
  • Agents vs. agentic systems vs. LLMs
  • Agency levels (e.g., workflows, multi-step agents)
Workflows
  • Prompt chaining
  • Routing
  • Parallelization (sectioning, voting)
  • Reflection
  • Orchestration-worker
Tools
  • Tool calling
  • Tool formatting
  • Tool execution
  • MCP
Multi-Step Agents
  • Planning autonomy
  • ReACT
  • Reflexion, ReWOO, etc.
  • Tree search for agents
Multi-Agent Systems (challenges, use-cases, A2A protocol)
Evaluation of agents

Project 4

Build "Deep Research" Capability with Web Search and Reasoning Models​

View attachment 295767
Reasoning and Thinking LLMs
  • Overview of reasoning models like OpenAI's "o" family and DeepSeek-R1
Inference-time Techniques
  • Inferece-time scaling
  • CoT prompting
  • Self-consistency
  • Sequential revision
  • Tree of Thoughts (ToT)
  • Search against a verifier
Training-time techniques
  • SFT on reasoning data (e.g., STaR)
  • Reinforcement learning with a verifier
  • Reward modeling (ORM, PRM)
  • Self-refinement
  • Internalizing search (e.g., Meta-CoT)

Project 5

Build a Multi-modal Generation Agent​

View attachment 295768
Overview of Image and Video Generation
  • VAE
  • GANs
  • Auto-regressive models
  • Diffusion models
Text-to-Image (T2I)
  • Data preparation
  • Diffusion architectures (U-Net, DiT)
  • Diffusion training (forward process, backward process)
  • Diffusion sampling
  • Evaluation (image quality, diversity, image-text alignment, IS, FID, and CLIP score)
Text-to-Video (T2V)
  • Latent-diffusion modeling (LDM) and compression networks
  • Data preparation (filtering, standardization, video latent caching)
  • DiT architecture for videos
  • Large-scale training challenges
  • T2V's overall system

Project 6

Capstone Project​

View attachment 295769
  • Choose your own idea
  • Build with techniques from the course
  • Get real-time feedback from the instructor as you build
  • Demo + feedback session


Link:
* Hidden text: cannot be quoted. *

View attachment 295763
hu
 

Create an account or login to comment

You must be a member in order to leave a comment

Create account

Create an account on our community. It's easy!

Log in

Already have an account? Log in here.

Tips
Recently searched:

Similar threads

Users who are viewing this thread

Top Bottom