cyrano

joined 2 years ago
[–] cyrano@lemmy.dbzer0.com 14 points 3 days ago (3 children)

But it is interesting that they are not using their own product to build it…..since “you don’t need developers anymore”

[–] cyrano@lemmy.dbzer0.com 4 points 3 days ago

Here are some points to consider:

  • Distribution: The focus isn't on distribution, as OpenAI has a larger user base.
  • Talent: The emphasis isn't on the team, since they specialize in UX rather than ML search.
  • Why not vibe coding it? Likely because the true value lies in the data rather than the user experience: the actions each user takes to resolve or build code.
[–] cyrano@lemmy.dbzer0.com 9 points 3 days ago (3 children)

What is the reference?

[–] cyrano@lemmy.dbzer0.com 8 points 4 days ago (2 children)

Yeah we only know about the chat from the DUI guy. Every body must have their side chat. Bessent for his Wall Street buddies, Sacks to push it to his all in VIP crypto group…. Etc. Seriously 😐

[–] cyrano@lemmy.dbzer0.com 32 points 4 days ago (4 children)

Hi Pete, it's your Uncle Bob. Could you please add me to the group chat for tariff announcements and trading tips? Thanks!

[–] cyrano@lemmy.dbzer0.com 47 points 4 days ago (1 children)

Withdrawing from the fund would be a significant own goal. If it were to do so, the US would lose all influence over the IMF’s policies and operations. More importantly, withdrawal would dramatically diminish the international role of the US dollar.

Hmm so probably going to be announced next week

[–] cyrano@lemmy.dbzer0.com 29 points 6 days ago (19 children)

But isn’t the domain already doing that?

[–] cyrano@lemmy.dbzer0.com 22 points 6 days ago

If you buy all the competition, you can set the price/rules.

[–] cyrano@lemmy.dbzer0.com 15 points 6 days ago

Verification wise there is already domain. But ultimately, it is too soon for the twitter exodus to get the blue check. All in all, this type of outrage is doomed to repeat with that type of central entity.

 

For now, the artificial intelligence tool named Neutron Enterprise is just meant to help workers at the plant navigate extensive technical reports and regulations — millions of pages of intricate documents from the Nuclear Regulatory Commission that go back decades — while they operate and maintain the facility. But Neutron Enterprise’s very existence opens the door to further use of AI at Diablo Canyon or other facilities — a possibility that has some lawmakers and AI experts calling for more guardrails.

13
The Llama 4 herd (ai.meta.com)
submitted 2 weeks ago* (last edited 2 weeks ago) by cyrano@lemmy.dbzer0.com to c/technology@lemmy.world
 
Llama 4 Models:
  - Both Llama 4 Scout and Llama 4 Maverick use a Mixture-of-Experts (MoE) design with 17B active parameters each.
  - They are natively multimodal: text + image input, text-only output.
  - Key achievements include industry-leading context lengths, strong coding/reasoning performance, and improved multilingual capabilities.
  - Knowledge cutoff: August 2024.

  Llama 4 Scout:
  - 17B active parameters, 16 experts, 109B total.
  - Fits on a single H100 GPU (INT4-quantized).
  - 10M token context window
  - Outperforms previous Llama releases on multimodal tasks while being more resource-friendly.
  - Employs iRoPE architecture for efficient long-context attention.
  - Tested with up to 8 images per prompt.

  Llama 4 Maverick:
  - 17B active parameters, 128 experts, 400B total.
  - 1M token context window.
  - Not single-GPU; runs on one H100 DGX host or can be distributed for greater efficiency.
  - Outperforms GPT-4o and Gemini 2.0 Flash on coding, reasoning, and multilingual tests at a competitive cost.
  - Maintains strong image understanding and grounded reasoning ability.

  Llama 4 Behemoth (Preview):
  - 288B active parameters, 16 experts, nearly 2T total.
  - Still in training; not yet released.
  - Exceeds GPT-4.5, Claude Sonnet 3.7, and Gemini 2.0 Pro on STEM benchmarks (e.g., MATH-500, GPQA Diamond).
  - Serves as the “teacher” model for Scout and Maverick via co-distillation.

  Misc:
  - MoE Architecture: Only 17B parameters activated per token, reducing inference cost.
  - Native Multimodality: Unified text + vision encoder, pre-trained on large-scale unlabeled data.
view more: ‹ prev next ›