Technology

34728 readers
123 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
2126
2127
 
 

cross-posted from: https://lemmy.world/post/708817

Visit TheBloke's HuggingFace page to see all of the new models in their SuperHOT glory.

SuperHOT models are LLMs who's LoRAs have been adapted to support a context length of 8,000 tokens!

For reference, this is x4 times the default amount of many LLMs (i.e. 2048 tokens). Even some of the newer ones can only reach a context length of 4096 tokens, half the amount of these SuperHOT models!

Here are a few that were released if you couldn't view his HuggingFace:

New GPTQ Models from TheBloke

  • airoboros (13B)
  • CAMEL (13B)
  • Chronos (13B)
  • Guanaco (13B & 33B)
  • Manticore (13B)
  • Minotaur (13B)
  • Nous Hermes (13B)
  • Pygmalion (13B)
  • Samantha (13B & 33B)
  • Snoozy (13B)
  • Tulu (13B & 33B)
  • Vicuna (13B & 33B)
  • WizardLM (13B)

We owe a tremendous thank you to TheBloke, who has enabled many of us in the community to interact with versions of Manticore, Nous Hermes, WizardLM and others running the remarkable 8k context length from SuperHOT.

Many of these are 13B models, which should be compatible with consumer grade GPUs. Try using Exllama or Oobabooga for testing out these new formats.

Shoutout to Kaikendev for the creation of SuperHOT. You can learn more about their work here.

If you enjoyed reading this, please consider subscribing to /c/FOSAI where I do my best to keep you in the know with the latest and greatest advancements regarding free open-source artificial intelligence.

2128
2129
2130
2131
 
 

cross-posted from: https://lemmy.world/post/699214

Hello everyone!

I'd like to share with you another new resource: NLP Cloud - a provider and platform aimed to help you streamline AI/LLM deployments for your business or project.

If you're considering a startup in AI, this is a valuable read and resource. Support these developers by visiting their site and checking out the platform for yourself.

NLP Cloud (Natural Language Processing Cloud)

GPT-3, GPT-4, ChatGPT, GPT-J, and generative models in general, are very powerful AI models. We're showing you here how to effectively use these models thanks to few-shot learning, also known as prompt engineering. Few-shot learning is like training/fine-tuning an AI model, by simply giving a couple of examples in your prompt.

GPT-3, GPT-4, And ChatGPT

GPT-3, GPT-4, and ChatGPT, released by OpenAI, are the most powerful AI model ever released for text understanding and text generation.

GPT-3 was trained on 175 billion parameters, which makes it extremely versatile and able to understanding pretty much anything! We do not know the number of parameters in GPT-4 but results are even more impressive.

You can do all sorts of things with these generative models like chatbots, content creation, entity extraction, classification, summarization, and much more. But it takes some practice and using them correctly might require a bit of work.

GPT-J, GPT-NeoX, And Dolphin

GPT-NeoX and GPT-J are both open-source Natural Language Processing models, created by, a collective of researchers working to open source AI (see EleutherAI's website).

GPT-J has 6 billion parameters and GPT-NeoX has 20 billion parameters, which makes them the most advanced open-source Natural Language Processing models as of this writing. They are direct alternatives to OpenAI's proprietary GPT-3 Curie.

These models are very versatile. They can be used for almost any Natural Language Processing use case: text generation, sentiment analysis, classification, machine translation,... and much more (see below). However using them effectively sometimes takes practice. Their response time (latency) might also be longer than more standard Natural Language Processing models.

GPT-J and GPT-NeoX are both available on the NLP Cloud API. On NLP Cloud you can also use Dolphin, an in-house advanced generative model that competes with ChatGPT, GPT-3, and even GPT-4. Below, we're showing you examples obtained using the GPT-J endpoint of NLP Cloud on GPU, with the Python client. If you want to copy paste the examples, please don't forget to add your own API token. In order to install the Python client, first run the following: pip install nlpcloud.

Few-Shot Learning

Few-shot learning is about helping a machine learning model make predictions thanks to only a couple of examples. No need to train a new model here: models à la GPT-3 and GPT-4 are so big that they can easily adapt to many contexts without being re-trained.

Giving only a few examples to the model does help it dramatically increase its accuracy.

In Natural Language Processing, the idea is to pass these examples along with your text input. See the examples below!

Also note that, if few-shot learning is not enough, you can also fine-tune GPT-3 on OpenAI's website and GPT-J and Dolphin on NLP Cloud so the models are perfectly tailored to your use case.

You can easily test few-shot learning on the NLP Cloud Playground, in the text generation section. Click here to try text generation on the Playground. Then simply use one of the examples showed below in this article and see for yourself.

If you use a model that understands natural human instructions like ChatGPT or ChatDolphin, you might not always have to use few-shot learning, but it is alway interesting to apply few-shot learning when possible in order to get the most advanced results. If you do not want to use few-shot learning, read our dedicated guide about how to use ChatGPT and ChatDolphin with simple instructions: see the article here.

In my opinion, I think this is a big highlight of this service:

Data Privacy And Security

NLP Cloud is HIPAA / GDPR / CCPA compliant, and working on the SOC 2 certification. We cannot see your data, we do not store your data, and we do not use your data to train our own AI models.

You can read the full page and article here. If you're still interested, consider checking out this other amazing resource detailing how to utilize chat-gpt alternatives.

2132
2133
2134
2135
 
 

cross-posted from: https://programming.dev/post/246221

Folks in this group would be well aware of various cloud load-balancers. Today I would like to introduce LoxiLB.

LoxiLB is a modern open source cloud-native load-balancer which uses goLang and eBPF tech provided by the Linux kernel. It's architecture and offerings makes it fast and flexible compared to others.

Check out this link to know what makes LoxiLB different from others:

Other resources:

Feel free to explore and give it a try!!!

2136
2137
 
 

Possibly one of the most important invention of the 20th century

2138
 
 

...Years ago, the web used to be a place where individuals made things. They made homepages, forums, and mailing lists, and a small bit of money with it. Then companies decided they could do things better. They created slick and feature-rich platforms and threw their doors open for anyone to join. They put boxes in front of us, and we filled those boxes with text and images, and people came to see the content of those boxes...

2139
 
 

would be cool if common folk could also be forgiven for "accidentally" deleting evidence

2140
2141
2142
2143
 
 

Hello. I bought a new computer recently, and the computer I'm replacing is still good, but it only works as a tablet now. I'm considering putting Linux on it, but would it be worth it or should I get rid of my old computer? Thanks!

2144
2145
2146
2147
2148
2149
2150
view more: ‹ prev next ›