send mail to support@abhimanu.com mentioning your email id and mobileno registered with us! if details not recieved
Resend Opt after 60 Sec.
By Loging in you agree to Terms of Services and Privacy Policy
Claim your free MCQ
Please specify
Sorry for the inconvenience but we’re performing some maintenance at the moment. Website can be slow during this phase..
Please verify your mobile number
Login not allowed, Please logout from existing browser
Please update your name
Subscribe to Notifications
Stay updated with the latest Current affairs and other important updates regarding video Lectures, Test Schedules, live sessions etc..
Your Free user account at abhipedia has been created.
Remember, success is a journey, not a destination. Stay motivated and keep moving forward!
Refer & Earn
Enquire Now
My Abhipedia Earning
Kindly Login to view your earning
Support
Type your modal answer and submitt for approval
Consider the following statements regarding DeepSeek AI models and its implications:
DeepSeek’s R1 model is designed to be more energy-efficient than traditional AI models by using fewer GPUs.
The projected annual energy consumption of DeepSeek’s AI infrastructure by 2027 is expected to match the electricity demand of Japan.
The “Mixture of Experts” approach in DeepSeek’s models enables cost-effective AI development by allowing specialized models to collaborate.
Which of the statements given above are correct?
1&2 only
2&3 only
1&3 only
All of the above
Only statements 1&3 are correct.
Owing to its optimal use of scarce resources, DeepSeek has been pitted against US AI powerhouse OpenAI, as it is widely known for building large language models. DeepSeek-V3, one of the first models unveiled by the company, earlier this month surpassed GPT-4o and Claude 3.5 Sonnet in numerous benchmarks.
Deepseek’s R1 model uses only 2,000 GPUs compared to OpenAI’s 16,000+, reducing electricity consumption.
Statement 3 is correct: DeepSeek-V3 stands out because of its architecture, known as Mixture-of-Experts (MOE).
The MOE models are like a team of specialist models working together to answer a question, instead of a single big model managing everything.
The DeepSeek-V3 model is trained on 14.8 trillion tokens, which includes large, high-quality datasets that offer the model greater understanding of language and task-specific capabilities.
Additionally, the model uses a new technique known as Multi-Head Latent Attention (MLA) to enhance efficiency and cut costs of training and deployment, allowing it to compete with some of the most advanced models of the day.
Hence option 2nd is correct.
By: Shubham Tiwari ProfileResourcesReport error
Access to prime resources
New Courses