Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

Fredtt3Β 
posted an update about 2 hours ago
mrfakenameΒ 
posted an update about 4 hours ago
kadirnarΒ 
posted an update about 4 hours ago
view post
Post
112
Midjourney + Custom SDXL-Lightning:
  • 1 reply
Β·
not-lainΒ 
posted an update about 6 hours ago
view post
Post
231
If you're a researcher or developing your own model πŸ‘€ you might need to take a look at huggingface's ModelHubMixin classes.
They are used to seamlessly integrate your AI model with huggingface and to save/ load your model easily πŸš€

1️⃣ make sure you're using the appropriate library version
pip install -qU "huggingface_hub>=0.22"

2️⃣ inherit from the appropriate class
from huggingface_hub import PyTorchModelHubMixin
from torch import nn

class MyModel(nn.Module,PyTorchModelHubMixin):
  def __init__(self, a, b):
    super().__init__()
    self.layer = nn.Linear(a,b)
  def forward(self,inputs):
    return self.layer(inputs)

first_model = MyModel(3,1)

4️⃣ push the model to the hub (or use save_pretrained method to save locally)
first_model.push_to_hub("not-lain/test")

5️⃣ Load and initialize the model from the hub using the original class
pretrained_model = MyModel.from_pretrained("not-lain/test")

Salama1429Β 
posted an update about 9 hours ago
view post
Post
356
πŸ“š Introducing the 101 Billion Arabic Words Dataset

🌐 Exciting Milestone in Arabic Language Technology! hashtag#NLP hashtag#ArabicLLM hashtag#LanguageModels

πŸš€ Why It Matters:
1. 🌟 Large Language Models (LLMs) have brought transformative changes, primarily in English. It's time for Arabic to shine!
2. 🎯 This project addresses the critical challenge of bias in Arabic LLMs due to reliance on translated datasets.

πŸ” Approach:
1. πŸ’ͺ Undertook a massive data mining initiative focusing exclusively on Arabic from Common Crawl WET files.
2. 🧹 Employed state-of-the-art cleaning and deduplication processes to maintain data quality and uniqueness.

πŸ“ˆ Impact:
1. πŸ† Created the largest Arabic dataset to date with 101 billion words.
2. πŸ“ Enables the development of Arabic LLMs that are linguistically and culturally accurate.
3. 🌍 Sets a global benchmark for future Arabic language research.


πŸ”— Paper: https://lnkd.in/dGAiaygn
πŸ”— Dataset: https://lnkd.in/dGTMe5QV

- πŸ”„ Share your thoughts and let's drive the future of Arabic NLP together!

hashtag#DataScience hashtag#MachineLearning hashtag#ArtificialIntelligence hashtag#Innovation hashtag#ArabicData
akhaliqΒ 
posted an update about 14 hours ago
view post
Post
805
Chameleon

Mixed-Modal Early-Fusion Foundation Models

Chameleon: Mixed-Modal Early-Fusion Foundation Models (2405.09818)

We present Chameleon, a family of early-fusion token-based mixed-modal models capable of understanding and generating images and text in any arbitrary sequence. We outline a stable training approach from inception, an alignment recipe, and an architectural parameterization tailored for the early-fusion, token-based, mixed-modal setting. The models are evaluated on a comprehensive range of tasks, including visual question answering, image captioning, text generation, image generation, and long-form mixed modal generation. Chameleon demonstrates broad and general capabilities, including state-of-the-art performance in image captioning tasks, outperforms Llama-2 in text-only tasks while being competitive with models such as Mixtral 8x7B and Gemini-Pro, and performs non-trivial image generation, all in a single model. It also matches or exceeds the performance of much larger models, including Gemini Pro and GPT-4V, according to human judgments on a new long-form mixed-modal generation evaluation, where either the prompt or outputs contain mixed sequences of both images and text. Chameleon marks a significant step forward in a unified modeling of full multimodal documents.
merveΒ 
posted an update about 15 hours ago
view post
Post
891
I got asked about PaliGemma's document understanding capabilities, so I built a Space that has all the PaliGemma fine-tuned doc models πŸ“„πŸ“ŠπŸ“–
merve/paligemma-doc
lamhieuΒ 
posted an update about 16 hours ago
view post
Post
645
πŸŽ‰ Happy to announce about the collection called "Blackhole". It is a black hole of high quality data in many fields, multilingual to train LLMs with SFT and DPO methods.
πŸ“¦ There are now over 30++ high-quality datasets available so you can start creating interesting models. It will be updated in the future, glad if it helps someone.

lamhieu/blackhole-66473b7feec034b4fb70818a
Ali-C137Β 
posted an update about 17 hours ago
eienmojikiΒ 
posted an update about 18 hours ago
view post
Post
544
πŸ‘€ Try new Anime Gen model - StarryXL

πŸͺ„ Starry XL has improved upon the Kohaku Epsilon model by targeting the specific styles of top Pixiv artists and expanding the character dataset to generate high-quality images.

✨ Starry is based on epsilon, and during training, the caption are overall close to Kohaku epsilon, so the overall usage is the same. Go to the model's page below to see in detail how to use it!

πŸ”Ž Resources:
- StarryXL v5.2 on Huggingface: eienmojiki/Starry-XL-v5.2
- Offical model page: https://civitai.com/models/448552?modelVersionId=499498
- Kohaku-XL Epsilon: https://civitai.com/models/399873?modelVersionId=445973

πŸ“ƒ Credits:
- Demo: @eienmojiki
- Model's author: kitarz