Commit 068fde83 authored by Fabian Ziegner's avatar Fabian Ziegner
Browse files

Update README.md

parent 34bc457f
......@@ -3,7 +3,27 @@ Workflow
1. Install required packages
2. Make sure a MongoDB instance is set up and runnning
3. Fill in your database credentials in .mongo.conf
3. Fill in your database credentials in .mongo.conf (see https://gitlab.com/MaxSchambach/mdbh)
4. Fill in your desired settings in config.yaml
5. Run zeroshot_sacred.py to start the experiment(s)
6. Run transform.py to get a dataframe containing experiment results
Configuration
-------------
- name: Name of the experiment
- user: Username of the database
- host: Address of the database
- port: Port of the database
- database: Name of the database
- auth: Name of the authentication database
- pw: Password of the database
- device: GPU to use (see nvidia-smi for device numbers)
- batch_size: Batch size
- representation: A huggingface model (see https://huggingface.co/models)
- multi_threshold: Score threshold to use (see mlmc.thresholds.thresholds_dict.keys()). Single-label datasets will always use max prediction.
- formatted: If formatting is set to True each class label is replaced by a more descriptive label. Furthermore, if the huggingface method is used the hypothesis is replaced as well.
- cut_sample: Trims the input text to the maximum input size of the language model.
- method: "huggingface" or "flair"
- whole_dataset: If True the entire dataset is used for classification.
- datasets: Datasets to use (see mlmc.data.register.keys())
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment