Skip to content

polifonia-project/llm4led

Repository files navigation

component-id type name description work-package pilot project resource release-date release-number release-link doi changelog licence copyright contributors related-components credits
llm4led
Software
Curation of documentary evidence, experiments with LED/GPT-4
This repository contains Python code for scraping data from LED (Listening Experience Database). The code processes the obtained data and uses the GPT-4 API to generate annotations from the submitted listening evidence.
WP4
CHILD
polifonia-project
24/04/2024
v1.0
Apache-2.0
Copyright (c) 2024 The Open University
Chukwudi "Festus" Uwasomba <https://github.com/cfestus>
reuses
documentary-evidence-benchmark

Curation of documentary evidence, experiments with LED/GPT-4

Description

This repository contains Python code for scraping data from LED (Listening Experience Database). The code processes the obtained data and uses the GPT-4 API to generate annotations from the submitted listening evidence.

Prerequisites

  • Python 3.x
  • OpenAI API key

To run the code

  1. Clone the Repository:
    git clone <repository-url>
    cd <repository-directory>
    
  2. Install Dependencies:
    pip install -r requirements.txt.
    
  3. API Configuration: Create a .env file in the root directory and add your OpenAI API key, that is your GPT-4 API.
    OPENAI_API_KEY="your_api_key_here"
    
  4. Usage: Once you have completed the setup, run the main script to generate annotations.
    run main.py
    

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Packages

No packages published

Contributors 2

  •  
  •