From 3d60a1c7c986950d6cb968a09a682c2a02d4a49b Mon Sep 17 00:00:00 2001
From: toyota <toyota@cl.uni-heidelberg.de>
Date: Fri, 30 Mar 2018 17:13:12 +0200
Subject: [PATCH] update README

---
 README.md     |  8 ++------
 lib/README.md | 13 +++++++++++++
 2 files changed, 15 insertions(+), 6 deletions(-)
 create mode 100644 lib/README.md

diff --git a/README.md b/README.md
index 633d5d1..2f016ce 100644
--- a/README.md
+++ b/README.md
@@ -4,8 +4,7 @@ This is an implementation of the CHERTOY system for the Word Sense Induction tas
 This project also contains an implementation of the baseline and 40 experiments with it.
 
 We experiment with language models, specific features and clustering algorithms based on the sense2vec and the sent2vec systems. 
-With a detailed research over 40 experiments we got an interesting insight on the effects of several
-feature combinations which resulted in our WSI system CHERTOY.
+After having performed 40 carefully designed experiments we obtained interesting insights on the effects of several feature combinations which resulted in our WSI system CHERTOY.
 
 The system creates semantic related clusters from the given snippets (the text fragments get back from the search engine) for each pre-defined ambiguous topic. 
 It makes the preprocessing of the input data, creates a language model using vector representations for each snippet with sense2vec and vector misture model (BOW representation with summarization for each snippet) and creates semantic clusters with the Mean Shift clustering algorithm.
@@ -111,10 +110,7 @@ The folder experiments contains an implementation of the baseline and 40 differe
 
 * lib
 
-The folder contains code for preprocessing Wikipedia Dataset to train own sent2vec models for the experiments, 
-preprocessed Wikipedia 2017 Dataset,
-two self-trained models of the Wikipedia 2017 Dataset, that we used in our experiments with sent2vec, 
-README file,
+The folder contains code for preprocessing Wikipedia Dataset to train own sent2vec models for the experiments and a README file. Our preprocessed Wikipedia 2017 dataset and two self-trained models of the Wikipedia 2017 dataset, that we used in our experiments with sent2vec, are provided on /proj/toyota on the server of the Institut.
 Other models that we used during our experiments can be found in sense2vec and sent2vec repositories.
 
 * experiments
diff --git a/lib/README.md b/lib/README.md
new file mode 100644
index 0000000..00e705e
--- /dev/null
+++ b/lib/README.md
@@ -0,0 +1,13 @@
+# CHERTOY - Creating language model with sent2vec
+
+This is an 
+
+## RUNNING INSTRUCTIONS
+
+### Input files:
+
+### Output files:
+
+### Create a folder structure:
+
+### RUN THE SYSTEM:
-- 
GitLab