John Smith PRO

John6666

AI & ML interests

None yet

Recent Activity

updated a dataset 11 minutes ago
John6666/flux1-backup-202411
New activity 13 minutes ago
John6666/flux-lora-the-explorer
updated a model 20 minutes ago
bhashwarsengupta/gemma-finance

Organizations

John6666's activity

reacted to singhsidhukuldeep's post with 👀 about 9 hours ago
view post
Post
345
It's always exciting to revisit Google's DCN paper—impractical but good!

Deep & Cross Network (DCN) - a groundbreaking approach to click-through rate prediction that's revolutionizing digital advertising!

Key Innovation:
DCN introduces a novel cross-network architecture that automatically learns feature interactions without manual engineering. What sets it apart is its ability to explicitly model bounded-degree feature crossings while maintaining the power of deep neural networks.

Technical Deep Dive:
- The architecture combines a cross network with a deep network in parallel.
- The cross network performs automatic feature crossing at each layer.
- The embedding layer transforms sparse categorical features into dense vectors.
- Cross layers use a unique formula that enables efficient high-degree polynomial feature interactions.
- Memory-efficient design with linear complexity O(d) in the input dimension.

Performance Highlights:
- Outperforms traditional DNN models with 60% less memory usage.
- Achieved 0.4419 logloss on the Criteo Display Ads dataset.
- Consistently performs better than state-of-the-art models like Deep Crossing and Factorization Machines.
- Exceptional performance on non-CTR tasks like Forest Covertype (97.40% accuracy).

Under the Hood:
- Uses embedding vectors of dimension 6 × (category cardinality)^1/4.
- Implements batch normalization and the Adam optimizer.
- The cross network depth determines the highest polynomial degree of feature interactions.
- An efficient projection mechanism reduces cubic computational cost to linear.
- Parameter sharing enables better generalization to unseen feature interactions.

Key Advantages:
1. No manual feature engineering required.
2. Explicit feature crossing at each layer.
3. Highly memory-efficient.
4. Scalable to web-scale data.
5. Robust performance across different domains.

Thoughts on how this could transform digital advertising?
reacted to m-ric's post with 👀 about 10 hours ago
view post
Post
424
🔍 Meta teams use a fine-tuned Llama model to fix production issues in seconds

One of Meta's engineering teams shared how they use a fine-tuned small Llama (Llama-2-7B, so not even a very recent model) to identify the root cause of production issues with 42% accuracy.

🤔 42%, is that not too low?
➡️ Usually, whenever there's an issue in production, engineers dive into recent code changes to find the offending commit. At Meta's scale (thousands of daily changes), this is like finding a needle in a haystack.
💡 So when the LLM-based suggestion is right, it cuts incident resolution time from hours to seconds!

How did they do it?

🔄 Two-step approach:
‣ Heuristics (code ownership, directory structure, runtime graphs) reduce thousands of potential changes to a manageable set
‣ Fine-tuned Llama 2 7B ranks the most likely culprits

🎓 Training pipeline:
‣ Continued pre-training on Meta's internal docs and wikis
‣ Supervised fine-tuning on past incident investigations
‣ Training data mimicked real-world constraints (2-20 potential changes per incident)

🔮 Now future developments await:
‣ Language models could handle more of the incident response workflow (runbooks, mitigation, post-mortems)
‣ Improvements in model reasoning should boost accuracy further

Read it in full 👉 https://www.tryparity.com/blog/how-meta-uses-llms-to-improve-incident-response
reacted to prithivMLmods's post with 👀 about 10 hours ago
view post
Post
1069
🍅 Glif App's Remixes feature allows you to slap a logo onto anything, seamlessly integrating the input image (logo) into various contexts. The result is stunning remixes that blend the input logo with generated images (img2img logo mapping) for incredible outcomes.

Check out Any Logo Anywhere remixes on Glif: [Glif Remixes](https://glif.app/glifs/cm3o7dfsd002610z48sz89yih/remixes)

🌐The browser extension enables thousands of Glif-based img2img workflows on any image you find online. Experience Glif Remix with WebAI: [Chrome Extension](https://chromewebstore.google.com/detail/glif-remix-the-web-with-a/abfbooehhdjcgmbmcpkcebcmpfnlingo)

.
.
.
🤗Have fun with the cool stuff !!
@prithivMLmods
reacted to etemiz's post with 👀 about 10 hours ago
view post
Post
590
if I host in hf spaces, can I interact with the app using an API?
  • 1 reply
·
reacted to vilarin's post with 🔥 about 10 hours ago
view post
Post
543
🏄‍♂️While browsing new models, I stumbled upon Lumiere from aixonlab. After testing it, I feel it has considerable potential. Keep up the good work!

Lumiere Alpha is a model focusing on improving realism without compromising prompt coherency or changing the composition completely from the original Flux.1-Dev model.

🦄 Model: aixonlab/flux.1-lumiere-alpha

🦖 Demo: vilarin/lumiere
  • 1 reply
·
reacted to jjokah's post with 👍 about 10 hours ago
view post
Post
381
Google's revamped Machine Learning Crash Course covers the recent advances in AI, with an increased focus on interactive learning.

📝 100+ exercises
🗂 12 modules
🕒 15 hours
📹 Video explainers of ML concepts
🌎 Real-world examples
📊 Interactive visualizations

Ref:
https://developers.google.com/machine-learning/crash-course
reacted to jsulz's post with 🔥 about 10 hours ago
view post
Post
1022
When the XetHub crew joined Hugging Face this fall, @erinys and I started brainstorming how to share our work to replace Git LFS on the Hub. Uploading and downloading large models and datasets takes precious time. That’s where our chunk-based approach comes in.

Instead of versioning files (like Git and Git LFS), we version variable-sized chunks of data. For the Hugging Face community, this means:

⏩ Only upload the chunks that changed.
🚀 Download just the updates, not the whole file.
🧠 We store your file as deduplicated chunks

In our benchmarks, we found that using CDC to store iterative model and dataset version led to transfer speedups of ~2x, but this isn’t just a performance boost. It’s a rethinking of how we manage models and datasets on the Hub.

We're planning on our new storage backend to the Hub in early 2025 - check out our blog to dive deeper, and let us know: how could this improve your workflows?

https://huggingface.co/blog/from-files-to-chunks
reacted to mgubri's post with 🔥 about 10 hours ago
view post
Post
413
🎉 We’re excited to announce, in collaboration with @kaleidophon , the release of the models from our Apricot 🍑 paper, "Apricot: Calibrating Large Language Models Using Their Generations Only," accepted at ACL 2024! Reproducibility is essential in science, and we've worked hard to make it as seamless as possible.
parameterlab/apricot-models-673d2cae40b6ff437a86f0bf
reacted to fdaudens's post with 👀 about 10 hours ago
view post
Post
455
🚀 DeepSeek just dropped DeepSeek-R1-Lite-Preview with “reasoning” capacity.

- Matches OpenAI o1-preview on AIME & MATH benchmarks.
- Transparent process output
- Open-source model to be released

Try it out: https://chat.deepseek.com/
reacted to rwightman's post with 🚀 about 10 hours ago
view post
Post
442
Want to validate some hparams or figure out what timm model to use before commiting to download or training with a large dataset? Try mini-imagenet: timm/mini-imagenet

I had this sitting on my drive and forgot where I pulled it together from. It's 100 classes of imagenet, 50k train and 10k val images (from ImageNet-1k train set), and 5k test images (from ImageNet-1k val set). 7.4GB instead of > 100GB for the full ImageNet-1k. This ver is not reduced resolution like some other 'mini' versions. Super easy to use with timm train/val scripts, checkout the dataset card.

I often check fine-tuning with even smaller datasets like:
* timm/resisc45
* timm/oxford-iiit-pet
But those are a bit small to train any modest size model w/o starting from pretrained weights.
reacted to openfree's post with 🚀 about 10 hours ago
view post
Post
501
MOUSE-I: Transform a Prompt into a Live Web Service
"From Prompt to Global Service in 60 Seconds"
The Future of Web Development
MOUSE-I revolutionizes web development by converting a single prompt into a fully functional, globally deployed web service through AI automation and enterprise-grade infrastructure.
⚡ Lightning-Fast Pipeline (60 Seconds)
1. AI Prompt Enhancement (5s)

Instant requirement analysis
Tech stack optimization
Development spec generation

2. Code Creation (49s)

Production-ready code
Responsive design
Performance-optimized

3. Live Rendering (1s)

Instant visualization
Real-time testing

4. Global Deployment (5s)

Vercel infrastructure
Global CDN
Automatic HTTPS

🎯 Key Differentiators

Instant Results: From idea to live URL in 60 seconds
Enterprise Quality: Production-grade code and infrastructure
Zero Configuration: No setup or technical knowledge required
40+ Templates: Ready-to-use solutions for games, dashboards, and apps

💫 Perfect For

Startups needing quick MVPs
Developers prototyping ideas
Non-technical founders building web services
Educators creating interactive tools

🚀 Get Started

Visit MOUSE-I Gallery
Enter your prompt
Get your live service in 60 seconds

💡 Connect

🌐 MOUSE-I Gallery
https://huggingface.co/spaces/VIDraft/mouse1


💬 discord.gg/openfreeai
📧 [email protected]
  • 1 reply
·
reacted to AlonzoLeeeooo's post with 🚀 about 10 hours ago
view post
Post
370
🎉 We are excited to announce our latest research on video editing - StableV2V!
💭 StableV2V aims to perform video editing with aligned shape consistency to user prompt, even if which might cause significant shape differences.
📚 Besides, we curate a testing benchmark, namely DAVIS-Edit, for video editing, comprising of both text-based and image-based applications.
🚀 We have open-sourced our paper, code, model weights, and DAVIS-Edit, which you may refer to more details of StableV2V from the following link:

- arXiv paper: https://arxiv.org/abs/2411.11045
- Project page: https://alonzoleeeooo.github.io/StableV2V/
- GitHub: https://github.com/AlonzoLeeeooo/StableV2V
- HuggingFace model repo: AlonzoLeeeooo/StableV2V
- HuggingFace dataset repo: AlonzoLeeeooo/DAVIS-Edit
replied to their post about 11 hours ago
view reply

In other words, there is a high possibility that the nuisance is being caused by that specific individual, or by someone trying to bring that specific individual down in a rare case...😅
Anyway, it explains why the scale is not large.

replied to their post about 13 hours ago
view reply

It's true. There are posts that don't do any harm, but the context doesn't make sense...🙀
Unlike Forum topics, it's harder to make Posts, but replies are easy. There are also more and more SPAM replies to Posts.
I don't think there's much point in tracking me, so I wonder if the bot is programmed to link to the actions of active users.
His icon is anime. When they automatically create an account, they seem to copy someone else's icon, so I don't know if he's an anime fan or not.

reacted to nyuuzyou's post with 🔥 about 14 hours ago
view post
Post
224
🎵 Introducing Tamago Music Dataset - nyuuzyou/tamago

A collection of 1,567 music tracks featuring:

- Complete metadata with audio files and cover artwork
- Rich track information including titles, descriptions, and genres
- User engagement metrics like play counts and reactions
- English language content from independent artists
- Released under Creative Commons Zero (CC0) license

Dataset structure includes:
- Track metadata (titles, descriptions, genres, tags)
- Associated media (audio files, cover images)
- Artist information and engagement metrics

Particularly valuable for:
- Music generation model training
- Cross-modal analysis
- Audio classification tasks
- Music style and genre analysis
replied to nroggendorff's post about 14 hours ago
replied to victor's post 1 day ago
view reply

QoL:

  • I myself is not really having any trouble, but if I had to say, I'd like the $18/month plan...🙄
    I don't know if this designation is an ethically appropriate way of referring to things in today's world, but it seems that people in developing countries with low prices and wages are having difficulties due to the recent huge increase in the size of AI models and the fact that the specifications for free CPU space have not changed. Even if we limit it to the active members of HF that I know, there are more than three people who do not have the income to subscribe to Pro. If they had the means, I'm sure they would contribute to the community.
    But if we were to give them unlimited CPU space, HF would go bankrupt...
    Is it possible to use a weaker version of the Pro subscription by applying for a system similar to the once-popular Community Grant? I think it would make a big difference if we could use even one strong space.
replied to victor's post 1 day ago
view reply

Can't this post be placed in a more prominent location?
I'm introducing it to the extent that it doesn't become SPAM, but after all, my sphere of activity is small...😅

replied to victor's post 1 day ago
view reply

https://huggingface.co/spaces/gaverfraxz/Weight_Comparator
I happened to find this space today, but if HF incorporates this kind of algorithm into the search engine simply, there will be some load on the server because it will have to read all the models.
The files read by from_pretrained() are determined by their names, so there is a way to simply add up their sizes, but that would be troublesome if models were quantized.
The size display in GGUF is built into the GUI, but I think that's possible because it's written in the header or something.

replied to victor's post 1 day ago