Clear Filters
Clear Filters

How can I train a neural network with information about academic courses (mostly text as input, 1 numeric) to classify whether the course is future proof?

3 views (last 30 days)
Hi guys,
The test database that I have is a has 6 input variables (5 in text, 1 numeric) and one (0 or 1) output variable saved in a table. These 6 input variables describe content of academic courses. The name, description, learning outcomes, level, #credits and department. In a training set of 93 courses I identified 20 courses that are FutureProof. This is a time intensive task. The rest of the database consist of ±2000 courses.
How can I train a neural network (if this is the right choice) on this training set to identify which of the other ±1900 courses can be classified as future proof and which cannot?
Thanks a lot in advance!

Answers (1)

Vishal Bhutani
Vishal Bhutani on 21 Sep 2018
By my understanding, you want to train a neural network which is having input data of both type numerical as well as nominal variables. The link attached will refer to the same issue you are facing:
Hope it helps.
  1 Comment
Christiaan Teeuwen
Christiaan Teeuwen on 24 Sep 2018
Edited: Christiaan Teeuwen on 24 Sep 2018
Dear Vishal,
Thanks for your answer!
If I understand the link that you posted correctly, they categorize in the following manner:
"Oil = 1; Water = 2; Slickwater = 3; Thebloodoftheinnocent = 4;
Fluid = Oil;"
Here the distiction is clear between the words that are transformed into categories. However, the input that I try to use consists of two instances with text varying from e.g. 10-100 words that are sentences and therefore not categorizable. The output variable has a 0 or 1 output.
I hope this clarifies the problem I am having. Would you know how to such a thing?
Kind regards, Christiaan

Sign in to comment.

Categories

Find more on Text Analytics Toolbox in Help Center and File Exchange

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!