Site Map - skip to main content

Hacker Public Radio

Your ideas, projects, opinions - podcasted.

New episodes every weekday Monday through Friday.
This page was generated by The HPR Robot at


hpr3953 :: Large language models and AI don't have any common sense

Learn how to load and run GPT-2 or Llama2 to test it with common sense questions.

<< First, < Previous, , Latest >>

Hosted by hobs on 2023-09-27 is flagged as Clean and is released under a CC-BY-SA license.
AI, GPT, GPT-2, Llama2, Hugging Face, Machine Learning, Deep Learning, Natural Language Processing. 1.
The show is available on the Internet Archive at: https://archive.org/details/hpr3953

Listen in ogg, spx, or mp3 format. Play now:

Duration: 00:18:10

general.

Hobson and Greg are working with volunteers to develop an open source AI that we call Qary (QA for question answering). We're adding plugins to support open source large language models (LLMs) like GPT-2 and Llama2. Here's how you can use LLMs in your own Python Programs.

  1. Create a Hugging Face account:
  1. Create and copy your access token:
  1. Create a .env file with your access token string:
echo "HUGGINGFACE_ACCESS_TOKEN=hf_..." >> .env
  1. Load the .env variables in your python script using dotenv package and os.environ:
  • TIP: Use os.environ to retrieve the dict of variable values rather than dotenv.load_values- Otherwise other environment variables that have been set by other shell scripts such as .bashrc will be ignored.
  • This confused us when we were getting our GitLab CI-CD pipeline working and deploying to Render.com.
  • Each of your cloud services will have different approaches to setting environment variables.
  • This token string can be passed as a keyword argument to most of the pipeline and model classes.
import dotenv
dotenv.load_dotenv()
import os
env = dict(os.environ)
token = env['HUGGINGFACE_ACCESS_TOKEN']
  1. Find the path and name for the model on Hugging Face hub you want to use:
  • search for "llama2" in the top search bar on huggingface.co/
  • TIP: don't hit enter at the end of your search, instead click on "See 3958 model results for llama2"
  • I clicked on meta-llama/Llama-2-7b-chat-hf to see the documentation
  1. On the documentation page for your model you may have to apply for a license if it's not really open source but business source like Meta does with its AI so you can't use their models to compete with them
  • Apply for a license to use Llama2 on ai.meta.com using the same e-mail you used for your Hugging Face account.
  1. Follow the instructions on huggingface.co to authenticate your python session
  • TIP: You'll need to use the kwarg use_auth_token in the AutoModel.from_pretrained or pipeline functions.
  • And it should be set to the token from your Hugging Face profile page. The hugging face documentation says to use the token kwarg, but that never worked for me.
from transformers import pipeline, set_seed
generator = pipeline('text-generation', model='openai-gpt')
q = "2+2="
responses = generator(
    q,
    max_length=10,
    num_return_sequences=10
    )
responses
[{'generated_text': '2+2= 2.2, 1.1 and'},
 {'generated_text': '2+2= 3336 miles. they'},
 {'generated_text': '2+2= 2, = 2 = 2'},
 {'generated_text': '2+2= 4 = 2 = 5 \n'},
 {'generated_text': '2+2= 0 ( 1 ) = ='},
 {'generated_text': '2+2= 6 times the speed of sound'},
 {'generated_text': '2+2= 2 times 5, 865'},
 {'generated_text': '2+2= 3 / 7 / 11 ='},
 {'generated_text': '2+2= 2 2 \n 2 of 2'},
 {'generated_text': '2+2= 1, 9 = 1,'}]

Here's the cow leg counting question:

    q = "There are 2 cows and 2 bulls, how many legs are there?"
    responses = generator(
        f"Question: {q}\nAnswer: ",
        max_length=30,
        num_return_sequences=10)
    answers = []
    for resp in responses:
        text = resp['generated_text']
        answers.append(text[text.find('Answer: ')+9:])
    answers
     'four. \n " let me see if i have this straight',
     'only 3. and three cows and 2 bulls are bigger than',
     '2, 2, 1, 2. \n " not yet',
     "one per cow, that's all there is. in fact",
     '30. and what am i? oh, yes, about',
     'one. the big, white bull that is bigger than 1',
     'three. they need to be introduced to the cow population before',
     "1. i don't know how many and where exactly ;",
     'no 2. 2. two bulls for 1 bull and 2',
     '1, there are 1.2, and 2, there are']

Comments

Subscribe to the comments RSS feed.

Comment #1 posted on 2023-10-01 06:55:46 by Mr Young

LLMs are great if you use them right

Great show. I've been using LLMs for work lately, and they are great at certain activities, as long as you don't expect them to act like humans with common sense. There are certain NLP tasks like document Q&A that were near impossible before LLMs that are a few lines of code now.

For a lay-person interacting with Bard, ChatGPT, etc. I recommend the following sites for understanding how to ask LLMs good prompts:

https://docs.google.com/presentation/d/17b_ocq-GL5lhV_bYSShzUgxL02mtWDoiw9xEroJ5m3Q/edit?pli=1#slide=id.gc6f83aa91_0_79

https://learnprompting.org/docs/intro https://www.promptingguide.ai/

Leave Comment

Note to Verbose Commenters
If you can't fit everything you want to say in the comment below then you really should record a response show instead.

Note to Spammers
All comments are moderated. All links are checked by humans. We strip out all html. Feel free to record a show about yourself, or your industry, or any other topic we may find interesting. We also check shows for spam :).

Provide feedback
Your Name/Handle:
Title:
Comment:
Anti Spam Question: What does the letter P in HPR stand for?
Are you a spammer?
What is the HOST_ID for the host of this show?
What does HPR mean to you?