Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

anakin87 
posted an update 1 day ago
view post
Post
1892
LLMs can leak their post-training data (RL included) 💧

New interesting paper on this topic from Google DeepMind: Extracting alignment data in open models (2510.18554)

It's known that Language Models memorize data that can be extracted via prompting.

In this paper, the authors investigate this aspect:
- using open models, where prompting can be fully customized by the user, including special tokens.
- focusing on open-source models like Olmo, where full training data is available.


📤 How do they extract data?

During post-training (like SFT), new tokens such as <|user|> are introduced.

The authors hypothesize prompting the model with these tokens can make it output its alignment data (remember Magpie?).

For example, for SFT, their extraction prompt is <|endoftext|><|user|>.


📏 Evaluating memorization

The authors compare each sampled example with the original data using vector search with embedding similarity.

They find that many outputs are semantically very similar to the original data, even if the exact words differ.

Traditional string-matching algorithms underestimate memorization by 10x.


🔁 What about RL?

Surprisingly, the same technique works to extract data from Reinforcement Learning (PPO/GRPO) phases.

This is counter-intuitive because the RL objective is not designed to increase sequence likelihoods (unlike SFT).

Practical limitation: in this case, extraction relies on using the initial part of the training prompt, which is not generally public.


📈 Are the extracted data effective for post-training?

Both in SFT and RL, the extracted data can be used to fine-tune models to similar performance to the originals.

The authors suggest that model distillation, where a stronger model is used to drive the training of a weaker one, may be a form of indirect training on the original dataset.

sequelbox 
posted an update 2 days ago
view post
Post
2337
NEW RELEASE: UML Generator is here!

- Our newest Experimental Reasoning release: create Unified Modeling Language diagrams to provide analysis and insight into your queries and situations!
- Multi-step reasoning reliably identifies diagram structure before a user response of XMI 2.5.1 code containing the UML diagram. Load the diagram into the UML tool of your choice!
- Trained in a variety of subjects for flexible analysis: software architecture, software development, business processes, systems engineering, data modeling, microservices, reverse engineering and more!

UML Generator available for multiple sizes of gpt-oss and Qwen 3, to provide increased flexibility to the user:
gpt-oss-120b: sequelbox/gpt-oss-120b-UML-Generator
gpt-oss-20b: sequelbox/gpt-oss-20b-UML-Generator
Qwen3-14B: sequelbox/Qwen3-14B-UML-Generator
Qwen3-4B-Thinking-2507: sequelbox/Qwen3-4B-Thinking-2507-UML-Generator

You can also get the UML Generator dataset, to train your own models to use UML Generator Format: sequelbox/UML-Generator-Dataset-DeepSeek-V3.2

Support our experimental open-source research efforts, models and datasets: sequelbox/SupportOpenSource

See our other Experimental Reasoning models: https://huggingface.co/collections/sequelbox/experimental-reasoning-models

with love,
allegra
unmodeled-tyler 
posted an update 2 days ago
view post
Post
1632
NEW RESEARCH MODEL: vanta-research/atom-v1-preview-4b

We are excited to share our Atom V1 4B Preview model! This fine tuned Gemma3 4B variant has a distinct, friendly, and exploratory persona - designed to help the user think and reflect.

Atom is trained to ask questions, use approachable, yet relatable analogies in ELI5-style explanations, and engage in deep, reflective conversation.

We plan to scale Atom's persona to larger architectures, and this iteration was created as part of that R&D.

Any and all feedback is always welcome as we continue to refine our approach.
ronantakizawa 
posted an update 1 day ago
view post
Post
2041
Introducing the Japanese honorifics dataset: a dataset with 137 sentences covering the three main keigo forms: 尊敬語 (Sonkeigo), 謙譲語 (Kenjōgo), and 丁寧語 (Teineigo). Each entry includes the base form, all three honorific transformations, and English translations for essential phrases in Japanese. This dataset is perfect for training and evaluating the Japanese skill level of LLMs.

#japanese #japanesedataset

ronantakizawa/japanese-honorifics
prithivMLmods 
posted an update 1 day ago
view post
Post
2258
Try the all-new trending Qwen-Image-Edit specialized adapter demos, including Photo-to-Anime, Light Restoration, Multi-Angle Edits, Relighting, and more — all in a single Hugging Face Space. Below is the demo link. 🤗🌠

⮞ Demo-Space: prithivMLmods/Qwen-Image-Edit-2509-LoRAs-Fast
⮞ How-to-Use: prithivMLmods/Qwen-Image-Edit-2509-LoRAs-Fast#2
⮞ Collection: https://huggingface.co/collections/prithivMLmods/image-generation-apps-collection

To know more about it, visit the app page or the respective model page!
  • 2 replies
·
wang12390 
posted an update 2 days ago
view post
Post
2258
AI Clothes Changer -Virtual Try On

Try on new outfits in your photos with the Miragic AI Clothes Changer. Instantly switch styles, test looks, or change clothes in images with just one click—no editing skills needed. It's fast, free, and fun to use!
https://miragic.ai/products/virtual-try-on
  • 1 reply
·
fd3ffff 
posted an update 2 days ago
view post
Post
1725
The methods for generating these interesting AI images:
1. Open https://imini.com/
2. Enter the instruction word.
3. Click "Generate Now".
4. Download, save or share on social media platforms.
mokady 
posted an update 3 days ago
TravisMuhlestein 
posted an update 3 days ago
view post
Post
1304
Building Smarter AI Agents: A Tool-Based Architecture for Modularity and Trust

Over the past year, our AI engineering team at GoDaddy has been rethinking how to make agent systems more modular, transparent, and production-ready. Instead of viewing an AI agent as a monolithic process, we’ve decomposed it into four core tools that separate decision-making from execution — a design that’s proving critical for scale and observability:

🧩 MemoryTool – maintains persistent context and user continuity
✅ CompletionTool – determines when a task is truly complete
💬 UserInteractionTool – manages clarifications, approvals, and confirmations
🔁 DelegationTool – enables agents to hand off tasks to other agents or humans

This approach makes every step of an agent’s workflow explicit, testable, and auditable, allowing us to scale AI systems in production with higher confidence. We see this as a step toward a more open, composable agent ecosystem — one where frameworks can interoperate and agents can build trust through transparency and version control.

Read the full write-up here → Building AI Agents at GoDaddy – An Agent’s Toolkit https://www.godaddy.com/resources/news/building-ai-agents-at-godaddy-an-agents-toolkit

We’d love to collaborate and exchange ideas with the community:

- How are you designing modular agent architectures?
- What design patterns or abstractions have helped you manage agent complexity?

Let’s build smarter, safer agents together.

#AI #Agents #Architecture #MachineLearning #OpenSource #AgentFrameworks #TrustInAI
wang12390 
posted an update 3 days ago
view post
Post
3419
How to Use the AI Image Generator
https://miragic.ai/products/image-generator
1. Describe Your Vision: Enter a text prompt describing what you want to create. Example: "A futuristic city skyline at sunset with glowing airships."
2. Select a Style: Choose an art style—Realistic, Anime, Painterly, Surreal, or Minimalist—to match your idea.
3. Generate and Refine: Click "Generate Image" and let the AI do its magic. Want to tweak it? Refine your prompt or try a new style.
4. Download and Share: Save your creation in high resolution or share it directly on social media.