20
мар
The easiest way to know the location of Oracle's trace and dump files is to use the server manager utility. As we see below, server manager can be used to display all of the dump parameters. These same parameters are used in the init.ora file to define the log and trace options. SVRMGR show parameters dump.
A database dump (also: SQL dump) contains a record of the table structure and/or the data from a database and is usually in the form of a list of SQL statements. A database dump is most often used for backing up a database so that its contents can be restored in the event of data loss. Corrupted databases can often be recovered by analysis of the dump. Database dumps are often published by free software and free content projects, to allow reuse or forking of the database.
See also[edit]
- SQLyog- MySQL GUI tool to generate Database dump
External links[edit]
- PostgreSQL dump backup methods, for PostgreSQL databases.
This Spring I worked with Brian Salter and Rich Jeffery to complete a semester long project using relational databases to predict upcoming movie outcomes. We wanted to do something with an exceedingly large amount of data, just for fun - and decided that movies were a good example of something with large amounts of data that would fit well in relations.
We came up with a motivation for our research:
Using existing data on past movies and available data on upcoming movies, can we accurately predict how well upcoming movies will do?
More specifically, using metrics of past movies such as cast, director, release date, rating, duration, subject matter, etc, can we accurately predict whether a movie will be 'good' or 'bad' (perhaps measure in ratings, box office sales, etc)? What patterns in these metrics would ensure a quality prediction?
For our data, we used what anyone would expect us to use: the Internet Movie Database. From the packaged text data dumps that IMDb provides, we used an open source Python project called IMDbpy to extract all of the information and store it in our database. The IMDb user review score is the value we used as the criteria for a good movie.
We evaluated the data concerning movies released between 2000 and 2010 in an attempt to train a perceptron (a classification model), and to use the perceptron to predict the IMDb user review score of new movies based on the chosen metrics.
TL;DR: We were able to get a classifier working with an accuracy rate of over 60%, and ~80% with the threshold rating increased. It's an interesting project that we are considering improving on this summer and building some sort of hosted web app. Brian has started a GitHub project called PUMA (Probably Über Movie Advice) where you can see our classifier in action.
Importing the Data
IMDb is an exceedingly large database – getting all of the data into a database for our analysis was a bit of a challenge. The methods for processing the data and the very database implementation itself had to be changed from our original plan due to the sheer amount of data being handled and stored.
To begin with, we considered using a RESTful web API called IMDB API to retrieve data. This plan was determined to be quite a hassle, as processing would take an extraordinary amount of time and bandwidth and IMDB API limits the amount of requests to 3,000 per hour. The only way to describe this would be ‘unfortunate’, as there are 4,115,950 entries in our table containing persons within our final database structure. That would be over 57 days of data processing for one of our 22 tables. Instead we discovered that IMDb distributes subsets of plain text data files from their website on their Alternative Interfaces page. The 49 unique data files require 6.75GB of plain text when uncompressed.
To process the enormous plain text files, we used an open source project called IMDbPY. The project is written entirely in Python (with a few lines of C!) and is released under the GPL license. It is primarily used as an API library for individual queries in Python, but also has a script (about 3,000 lines of code) for processing of IMDb’s plain text files. The actual processing of data was time consuming and made us appreciate the intricacies and design choices of database structure and implementation. The many different retries in processing data led to many hours of downtime while processing occurred. These retries were largely due to some Python specific dependencies that needed to be fulfilled, but also due to some design choices that were forced upon us due to the nature of the data. The original proposal called for using Oracle database to house the data. As IMDbPY is open source, it was unsurprising to find that there was no support for Oracle (only SQLite, Postgres, and MySQL). It was proposed to process the data to SQLite, export to SQL statements, and import to Oracle. This option failed for two reasons.
First, although SQLite has large theoretical limits on database size (these limits are primarily determined by machine-specific computational limits more than anything else), we found processing usually crashed the local machine. After many attempts to export the database to SQL using various methods, we eventually gave up on SQLite and looked for alternative methods, which brought us to MySQL.
After successfully importing to MySQL, we encountered a second problem: Oracle has a table size limit of 4GB. The IMDb data exceeds this, and in order to properly join and create intermediate relations for our analysis, we needed more space. This was a sufficient enough reason for our project to remain in MySQL rather than transferring it to Oracle.
It took about 1.5 hours in the final run to process on a decent machine (AMD quad-core, 8GB DDR3). Lesson learned: Big data requires thoughtful planning and knowledge to set up and maintain effectively.
Understanding the Structure
Upon seeing the final product of our data import, the format of our database was slightly daunting, not to mention the sheer amount of ‘useless’ data – data that was inconsequential to our analysis such as cinematographers, Disney Channel television shows, and aliases of makeup artists from documentaries. Every last detail about any one ‘thing’ seemed to be hidden away in its own separate table. Many of the columns we needed to use for analysis were stored in three or four tables, due to the normal form of the database as a whole. The spread of the data was an attempt to normalize the data, as redundancy on a database this size would be incredibly wasteful and hard to maintain.
Additionally, some information was stored in the same table, but different rows, requiring some finesse with our queries. For example, budget was stored in movie_info with an info_type_id of 105 (this is the key that identifies which type of information is in the row), and release date was stored in the same table with an info_type_id of 16. We also wanted to limit our analysis to theatrical release which required a join with the title table (using kind_id which is a key for identifying the type of release a movie is). A greatly condensed map of the tables that we used is below. Due to the massive size of each table, we were forced to limit our analysis to about a decade’s worth of data, considering only theatrical releases, and eliminating low budget films.
What if we just wanted to decide how good an actor is by examining the average scores of his movies?
A [very icky] nested query for this question looks like:
Running this query takes upwards of 10 minutes and tells us that the late Rod Steiger has an average review score of 6.2.
Trimming the Data
As you can guess from the above query, one of the largest immediate problems we encountered when we began our attempt to classify the data was the time for analysis, mostly due to the size of each row vector. Each row represented a movie, and each column represented an actor who was possibly in the film. Rather than store a movie as a collection of person_ids (which could potentially affect analysis) we stored a movie as a series of binary variables. Each actor has a person_id which represented the column index. In each column was a binary variable which indicated whether they were in the movie (1) or not (0). There have been (according to the data on IMDb) ~2.7 million actors since the dawn of time. Thus, each row had ~2.7 million columns. This made analysis incredibly slow.
To increase the speed we had to remove some useless data. Most noticeable was the fact that the overwhelming majority of columns represented actors who starred in no movies during the ten year span we were looking at, making the entire matrix very sparse. We therefore had to do some pruning to limit how much data we were processing.
Their relationship is a metaphor of how a child interacts with the big world and how an adult can aid with this difficult task. Masha and the bear in hindi download. Masha is a restless little girl who can hardly stand still due to her endless energy - everything she comes into contact with must be tried on! She's friendly beyond description and treats everyone like she's known them for ages.
We next noticed two strange phenomena in the data. First, there were hundreds (if not thousands) of records for independent films, which generally starred actors with little weight that had never been in a feature film. To correct this, we attempted to filter the records further by budget, however budget was stored as a character string, so direct comparison to an arbitrary minimum budget wasn’t possible. Instead we simply cut out all movies whose budget string was less than ten characters ($1,000,000). This removed most independent films. Second we noticed several movies appearing multiple times, which was causing them to appear multiple times in the actors database, which meant the thirty actors in a movie might appear three or four times for that same movie. For this, we simply ran several select distinct queries on the data to remove these duplicates as it seemed the only difference between them was release date. Maybe it was different geographic releases of the same film?
Once this was complete, we had managed to prune ~2.66 million columns off the data, and several thousand rows. This was met with a noticeable increase in processing speed when it came time to analyze the data again.
Predicting the Score
Once we had pulled all of the relevant data, we needed to train our classifiers. First we trained a perceptron to analyze the data. A perceptron is an algorithm that takes in a dataset and uses binary outputs to make predictions based on a linear prediction function. Because of this we had to arbitrarily choose a threshold by which to separate our data. We ran the perceptron using two different IMDB review scores as our metric for a ‘good’ movie. We first looked at 6.5 stars then at 7 stars to evaluate the accuracy of the perceptron. The perceptron works by computing a weight vector.
If the dot product of the weight vector and the feature vector is greater than some threshold (0 in this case) we classify it as 1, else we classify as 0. The goal of the training phase of the perceptron algorithm is to learn what the weight vector should be to minimize training error (number of misclassified points / total number of points). This algorithm is guaranteed to converge on a set of weights which correctly classify all training points, however there is no time limit for this and as such our algorithm was not allowed to converge.
Secondly we attempted to train a naïve Bayes classifier. This classifier works by using Bayes’s theorem to compute the probability of a particular class given a set of features. This is effectively the same system used in spam detection (P(Spam bunchawords)). The training of this model was the calculation of the probabilities P(Fi C) where Fi is a particular feature (1 or 0) and C is a particular class (1 or 0).
Once both models were trained, we calculated prediction performance for both models. For the perceptron, we simply compute a set of dot products and compare the result according to the above formula. We counted the number of misclassified points and used this to determine error. For Naïve Bayes we computed the formula below for each point:
In the 6.5 trials we tested only for the probability of the movie being good given the actor vector. For 7 we tested both P(Good actors) and P(Bad actors). Our results:
So we're on the right track! We were able to accurately predict whether a movie would be good or bad over 60% of the time.
We believe the error associated with the perceptron (testing error) is not a fully accurate measure of the system being better at 7 than 6.5. Because error is a function of misclassified points, and the probability of misclassification decreases as the boundary between good and bad increases, it becomes easier to predict as the boundary increases. That is, as we increase the boundary from 6.5 to 7, there are many more movies which count as bad than good. This results in it being easier to classify overall without the perceptron needing to be more intelligent.
Moving Forward
In the process of handling and manipulating the data, we had to make several large concessions. First, we had to limit our analysis to ten years to limit the matrix size. It is possible we could get better results looking through fifteen, twenty or even possibly thirty years of data. Secondly, we had to limit the number of features we looked at. Specifically, we only looked at actors within movies, but ignored features like writer, director, budget, time of year, rating and genre. We would like to use these additional metrics to learn better models.
Another large concession we had to make early on was to limit the complexity of our classifier. We had originally intended to attempt to predict actual star ratings rather than just good or bad based on an arbitrary value. This requires a multiclass perceptron which we haven't had time to build. We would also like to add layers to the perceptron, which would improve its ability to learn.
Credits
Much of this post has been snippets and excerpts of our group paper on the subject, and so due credit should be given to Rich and Brian. You can find Brian's site here, and don't forget to check out the continuation of this project on GitHub.
The easiest way to know the location of Oracle's trace and dump files is to use the server manager utility. As we see below, server manager can be used to display all of the dump parameters. These same parameters are used in the init.ora file to define the log and trace options. SVRMGR show parameters dump.
A database dump (also: SQL dump) contains a record of the table structure and/or the data from a database and is usually in the form of a list of SQL statements. A database dump is most often used for backing up a database so that its contents can be restored in the event of data loss. Corrupted databases can often be recovered by analysis of the dump. Database dumps are often published by free software and free content projects, to allow reuse or forking of the database.
See also[edit]
- SQLyog- MySQL GUI tool to generate Database dump
External links[edit]
- PostgreSQL dump backup methods, for PostgreSQL databases.
This Spring I worked with Brian Salter and Rich Jeffery to complete a semester long project using relational databases to predict upcoming movie outcomes. We wanted to do something with an exceedingly large amount of data, just for fun - and decided that movies were a good example of something with large amounts of data that would fit well in relations.
We came up with a motivation for our research:
Using existing data on past movies and available data on upcoming movies, can we accurately predict how well upcoming movies will do?
More specifically, using metrics of past movies such as cast, director, release date, rating, duration, subject matter, etc, can we accurately predict whether a movie will be 'good' or 'bad' (perhaps measure in ratings, box office sales, etc)? What patterns in these metrics would ensure a quality prediction?
For our data, we used what anyone would expect us to use: the Internet Movie Database. From the packaged text data dumps that IMDb provides, we used an open source Python project called IMDbpy to extract all of the information and store it in our database. The IMDb user review score is the value we used as the criteria for a good movie.
We evaluated the data concerning movies released between 2000 and 2010 in an attempt to train a perceptron (a classification model), and to use the perceptron to predict the IMDb user review score of new movies based on the chosen metrics.
TL;DR: We were able to get a classifier working with an accuracy rate of over 60%, and ~80% with the threshold rating increased. It's an interesting project that we are considering improving on this summer and building some sort of hosted web app. Brian has started a GitHub project called PUMA (Probably Über Movie Advice) where you can see our classifier in action.
Importing the Data
IMDb is an exceedingly large database – getting all of the data into a database for our analysis was a bit of a challenge. The methods for processing the data and the very database implementation itself had to be changed from our original plan due to the sheer amount of data being handled and stored.
To begin with, we considered using a RESTful web API called IMDB API to retrieve data. This plan was determined to be quite a hassle, as processing would take an extraordinary amount of time and bandwidth and IMDB API limits the amount of requests to 3,000 per hour. The only way to describe this would be ‘unfortunate’, as there are 4,115,950 entries in our table containing persons within our final database structure. That would be over 57 days of data processing for one of our 22 tables. Instead we discovered that IMDb distributes subsets of plain text data files from their website on their Alternative Interfaces page. The 49 unique data files require 6.75GB of plain text when uncompressed.
To process the enormous plain text files, we used an open source project called IMDbPY. The project is written entirely in Python (with a few lines of C!) and is released under the GPL license. It is primarily used as an API library for individual queries in Python, but also has a script (about 3,000 lines of code) for processing of IMDb’s plain text files. The actual processing of data was time consuming and made us appreciate the intricacies and design choices of database structure and implementation. The many different retries in processing data led to many hours of downtime while processing occurred. These retries were largely due to some Python specific dependencies that needed to be fulfilled, but also due to some design choices that were forced upon us due to the nature of the data. The original proposal called for using Oracle database to house the data. As IMDbPY is open source, it was unsurprising to find that there was no support for Oracle (only SQLite, Postgres, and MySQL). It was proposed to process the data to SQLite, export to SQL statements, and import to Oracle. This option failed for two reasons.
First, although SQLite has large theoretical limits on database size (these limits are primarily determined by machine-specific computational limits more than anything else), we found processing usually crashed the local machine. After many attempts to export the database to SQL using various methods, we eventually gave up on SQLite and looked for alternative methods, which brought us to MySQL.
After successfully importing to MySQL, we encountered a second problem: Oracle has a table size limit of 4GB. The IMDb data exceeds this, and in order to properly join and create intermediate relations for our analysis, we needed more space. This was a sufficient enough reason for our project to remain in MySQL rather than transferring it to Oracle.
It took about 1.5 hours in the final run to process on a decent machine (AMD quad-core, 8GB DDR3). Lesson learned: Big data requires thoughtful planning and knowledge to set up and maintain effectively.
Understanding the Structure
Upon seeing the final product of our data import, the format of our database was slightly daunting, not to mention the sheer amount of ‘useless’ data – data that was inconsequential to our analysis such as cinematographers, Disney Channel television shows, and aliases of makeup artists from documentaries. Every last detail about any one ‘thing’ seemed to be hidden away in its own separate table. Many of the columns we needed to use for analysis were stored in three or four tables, due to the normal form of the database as a whole. The spread of the data was an attempt to normalize the data, as redundancy on a database this size would be incredibly wasteful and hard to maintain.
Additionally, some information was stored in the same table, but different rows, requiring some finesse with our queries. For example, budget was stored in movie_info with an info_type_id of 105 (this is the key that identifies which type of information is in the row), and release date was stored in the same table with an info_type_id of 16. We also wanted to limit our analysis to theatrical release which required a join with the title table (using kind_id which is a key for identifying the type of release a movie is). A greatly condensed map of the tables that we used is below. Due to the massive size of each table, we were forced to limit our analysis to about a decade’s worth of data, considering only theatrical releases, and eliminating low budget films.
What if we just wanted to decide how good an actor is by examining the average scores of his movies?
A [very icky] nested query for this question looks like:
Running this query takes upwards of 10 minutes and tells us that the late Rod Steiger has an average review score of 6.2.
Trimming the Data
As you can guess from the above query, one of the largest immediate problems we encountered when we began our attempt to classify the data was the time for analysis, mostly due to the size of each row vector. Each row represented a movie, and each column represented an actor who was possibly in the film. Rather than store a movie as a collection of person_ids (which could potentially affect analysis) we stored a movie as a series of binary variables. Each actor has a person_id which represented the column index. In each column was a binary variable which indicated whether they were in the movie (1) or not (0). There have been (according to the data on IMDb) ~2.7 million actors since the dawn of time. Thus, each row had ~2.7 million columns. This made analysis incredibly slow.
To increase the speed we had to remove some useless data. Most noticeable was the fact that the overwhelming majority of columns represented actors who starred in no movies during the ten year span we were looking at, making the entire matrix very sparse. We therefore had to do some pruning to limit how much data we were processing.
Their relationship is a metaphor of how a child interacts with the big world and how an adult can aid with this difficult task. Masha and the bear in hindi download. Masha is a restless little girl who can hardly stand still due to her endless energy - everything she comes into contact with must be tried on! She's friendly beyond description and treats everyone like she's known them for ages.
We next noticed two strange phenomena in the data. First, there were hundreds (if not thousands) of records for independent films, which generally starred actors with little weight that had never been in a feature film. To correct this, we attempted to filter the records further by budget, however budget was stored as a character string, so direct comparison to an arbitrary minimum budget wasn’t possible. Instead we simply cut out all movies whose budget string was less than ten characters ($1,000,000). This removed most independent films. Second we noticed several movies appearing multiple times, which was causing them to appear multiple times in the actors database, which meant the thirty actors in a movie might appear three or four times for that same movie. For this, we simply ran several select distinct queries on the data to remove these duplicates as it seemed the only difference between them was release date. Maybe it was different geographic releases of the same film?
Once this was complete, we had managed to prune ~2.66 million columns off the data, and several thousand rows. This was met with a noticeable increase in processing speed when it came time to analyze the data again.
Predicting the Score
Once we had pulled all of the relevant data, we needed to train our classifiers. First we trained a perceptron to analyze the data. A perceptron is an algorithm that takes in a dataset and uses binary outputs to make predictions based on a linear prediction function. Because of this we had to arbitrarily choose a threshold by which to separate our data. We ran the perceptron using two different IMDB review scores as our metric for a ‘good’ movie. We first looked at 6.5 stars then at 7 stars to evaluate the accuracy of the perceptron. The perceptron works by computing a weight vector.
If the dot product of the weight vector and the feature vector is greater than some threshold (0 in this case) we classify it as 1, else we classify as 0. The goal of the training phase of the perceptron algorithm is to learn what the weight vector should be to minimize training error (number of misclassified points / total number of points). This algorithm is guaranteed to converge on a set of weights which correctly classify all training points, however there is no time limit for this and as such our algorithm was not allowed to converge.
Secondly we attempted to train a naïve Bayes classifier. This classifier works by using Bayes’s theorem to compute the probability of a particular class given a set of features. This is effectively the same system used in spam detection (P(Spam bunchawords)). The training of this model was the calculation of the probabilities P(Fi C) where Fi is a particular feature (1 or 0) and C is a particular class (1 or 0).
Once both models were trained, we calculated prediction performance for both models. For the perceptron, we simply compute a set of dot products and compare the result according to the above formula. We counted the number of misclassified points and used this to determine error. For Naïve Bayes we computed the formula below for each point:
In the 6.5 trials we tested only for the probability of the movie being good given the actor vector. For 7 we tested both P(Good actors) and P(Bad actors). Our results:
So we're on the right track! We were able to accurately predict whether a movie would be good or bad over 60% of the time.
We believe the error associated with the perceptron (testing error) is not a fully accurate measure of the system being better at 7 than 6.5. Because error is a function of misclassified points, and the probability of misclassification decreases as the boundary between good and bad increases, it becomes easier to predict as the boundary increases. That is, as we increase the boundary from 6.5 to 7, there are many more movies which count as bad than good. This results in it being easier to classify overall without the perceptron needing to be more intelligent.
Moving Forward
In the process of handling and manipulating the data, we had to make several large concessions. First, we had to limit our analysis to ten years to limit the matrix size. It is possible we could get better results looking through fifteen, twenty or even possibly thirty years of data. Secondly, we had to limit the number of features we looked at. Specifically, we only looked at actors within movies, but ignored features like writer, director, budget, time of year, rating and genre. We would like to use these additional metrics to learn better models.
Another large concession we had to make early on was to limit the complexity of our classifier. We had originally intended to attempt to predict actual star ratings rather than just good or bad based on an arbitrary value. This requires a multiclass perceptron which we haven't had time to build. We would also like to add layers to the perceptron, which would improve its ability to learn.
Credits
Much of this post has been snippets and excerpts of our group paper on the subject, and so due credit should be given to Rich and Brian. You can find Brian's site here, and don't forget to check out the continuation of this project on GitHub.
...">Imdb Database Dump Files Oracle(20.03.2020)The easiest way to know the location of Oracle's trace and dump files is to use the server manager utility. As we see below, server manager can be used to display all of the dump parameters. These same parameters are used in the init.ora file to define the log and trace options. SVRMGR show parameters dump.
A database dump (also: SQL dump) contains a record of the table structure and/or the data from a database and is usually in the form of a list of SQL statements. A database dump is most often used for backing up a database so that its contents can be restored in the event of data loss. Corrupted databases can often be recovered by analysis of the dump. Database dumps are often published by free software and free content projects, to allow reuse or forking of the database.
See also[edit]
- SQLyog- MySQL GUI tool to generate Database dump
External links[edit]
- PostgreSQL dump backup methods, for PostgreSQL databases.
This Spring I worked with Brian Salter and Rich Jeffery to complete a semester long project using relational databases to predict upcoming movie outcomes. We wanted to do something with an exceedingly large amount of data, just for fun - and decided that movies were a good example of something with large amounts of data that would fit well in relations.
We came up with a motivation for our research:
Using existing data on past movies and available data on upcoming movies, can we accurately predict how well upcoming movies will do?
More specifically, using metrics of past movies such as cast, director, release date, rating, duration, subject matter, etc, can we accurately predict whether a movie will be 'good' or 'bad' (perhaps measure in ratings, box office sales, etc)? What patterns in these metrics would ensure a quality prediction?
For our data, we used what anyone would expect us to use: the Internet Movie Database. From the packaged text data dumps that IMDb provides, we used an open source Python project called IMDbpy to extract all of the information and store it in our database. The IMDb user review score is the value we used as the criteria for a good movie.
We evaluated the data concerning movies released between 2000 and 2010 in an attempt to train a perceptron (a classification model), and to use the perceptron to predict the IMDb user review score of new movies based on the chosen metrics.
TL;DR: We were able to get a classifier working with an accuracy rate of over 60%, and ~80% with the threshold rating increased. It's an interesting project that we are considering improving on this summer and building some sort of hosted web app. Brian has started a GitHub project called PUMA (Probably Über Movie Advice) where you can see our classifier in action.
Importing the Data
IMDb is an exceedingly large database – getting all of the data into a database for our analysis was a bit of a challenge. The methods for processing the data and the very database implementation itself had to be changed from our original plan due to the sheer amount of data being handled and stored.
To begin with, we considered using a RESTful web API called IMDB API to retrieve data. This plan was determined to be quite a hassle, as processing would take an extraordinary amount of time and bandwidth and IMDB API limits the amount of requests to 3,000 per hour. The only way to describe this would be ‘unfortunate’, as there are 4,115,950 entries in our table containing persons within our final database structure. That would be over 57 days of data processing for one of our 22 tables. Instead we discovered that IMDb distributes subsets of plain text data files from their website on their Alternative Interfaces page. The 49 unique data files require 6.75GB of plain text when uncompressed.
To process the enormous plain text files, we used an open source project called IMDbPY. The project is written entirely in Python (with a few lines of C!) and is released under the GPL license. It is primarily used as an API library for individual queries in Python, but also has a script (about 3,000 lines of code) for processing of IMDb’s plain text files. The actual processing of data was time consuming and made us appreciate the intricacies and design choices of database structure and implementation. The many different retries in processing data led to many hours of downtime while processing occurred. These retries were largely due to some Python specific dependencies that needed to be fulfilled, but also due to some design choices that were forced upon us due to the nature of the data. The original proposal called for using Oracle database to house the data. As IMDbPY is open source, it was unsurprising to find that there was no support for Oracle (only SQLite, Postgres, and MySQL). It was proposed to process the data to SQLite, export to SQL statements, and import to Oracle. This option failed for two reasons.
First, although SQLite has large theoretical limits on database size (these limits are primarily determined by machine-specific computational limits more than anything else), we found processing usually crashed the local machine. After many attempts to export the database to SQL using various methods, we eventually gave up on SQLite and looked for alternative methods, which brought us to MySQL.
After successfully importing to MySQL, we encountered a second problem: Oracle has a table size limit of 4GB. The IMDb data exceeds this, and in order to properly join and create intermediate relations for our analysis, we needed more space. This was a sufficient enough reason for our project to remain in MySQL rather than transferring it to Oracle.
It took about 1.5 hours in the final run to process on a decent machine (AMD quad-core, 8GB DDR3). Lesson learned: Big data requires thoughtful planning and knowledge to set up and maintain effectively.
Understanding the Structure
Upon seeing the final product of our data import, the format of our database was slightly daunting, not to mention the sheer amount of ‘useless’ data – data that was inconsequential to our analysis such as cinematographers, Disney Channel television shows, and aliases of makeup artists from documentaries. Every last detail about any one ‘thing’ seemed to be hidden away in its own separate table. Many of the columns we needed to use for analysis were stored in three or four tables, due to the normal form of the database as a whole. The spread of the data was an attempt to normalize the data, as redundancy on a database this size would be incredibly wasteful and hard to maintain.
Additionally, some information was stored in the same table, but different rows, requiring some finesse with our queries. For example, budget was stored in movie_info with an info_type_id of 105 (this is the key that identifies which type of information is in the row), and release date was stored in the same table with an info_type_id of 16. We also wanted to limit our analysis to theatrical release which required a join with the title table (using kind_id which is a key for identifying the type of release a movie is). A greatly condensed map of the tables that we used is below. Due to the massive size of each table, we were forced to limit our analysis to about a decade’s worth of data, considering only theatrical releases, and eliminating low budget films.
What if we just wanted to decide how good an actor is by examining the average scores of his movies?
A [very icky] nested query for this question looks like:
Running this query takes upwards of 10 minutes and tells us that the late Rod Steiger has an average review score of 6.2.
Trimming the Data
As you can guess from the above query, one of the largest immediate problems we encountered when we began our attempt to classify the data was the time for analysis, mostly due to the size of each row vector. Each row represented a movie, and each column represented an actor who was possibly in the film. Rather than store a movie as a collection of person_ids (which could potentially affect analysis) we stored a movie as a series of binary variables. Each actor has a person_id which represented the column index. In each column was a binary variable which indicated whether they were in the movie (1) or not (0). There have been (according to the data on IMDb) ~2.7 million actors since the dawn of time. Thus, each row had ~2.7 million columns. This made analysis incredibly slow.
To increase the speed we had to remove some useless data. Most noticeable was the fact that the overwhelming majority of columns represented actors who starred in no movies during the ten year span we were looking at, making the entire matrix very sparse. We therefore had to do some pruning to limit how much data we were processing.
Their relationship is a metaphor of how a child interacts with the big world and how an adult can aid with this difficult task. Masha and the bear in hindi download. Masha is a restless little girl who can hardly stand still due to her endless energy - everything she comes into contact with must be tried on! She's friendly beyond description and treats everyone like she's known them for ages.
We next noticed two strange phenomena in the data. First, there were hundreds (if not thousands) of records for independent films, which generally starred actors with little weight that had never been in a feature film. To correct this, we attempted to filter the records further by budget, however budget was stored as a character string, so direct comparison to an arbitrary minimum budget wasn’t possible. Instead we simply cut out all movies whose budget string was less than ten characters ($1,000,000). This removed most independent films. Second we noticed several movies appearing multiple times, which was causing them to appear multiple times in the actors database, which meant the thirty actors in a movie might appear three or four times for that same movie. For this, we simply ran several select distinct queries on the data to remove these duplicates as it seemed the only difference between them was release date. Maybe it was different geographic releases of the same film?
Once this was complete, we had managed to prune ~2.66 million columns off the data, and several thousand rows. This was met with a noticeable increase in processing speed when it came time to analyze the data again.
Predicting the Score
Once we had pulled all of the relevant data, we needed to train our classifiers. First we trained a perceptron to analyze the data. A perceptron is an algorithm that takes in a dataset and uses binary outputs to make predictions based on a linear prediction function. Because of this we had to arbitrarily choose a threshold by which to separate our data. We ran the perceptron using two different IMDB review scores as our metric for a ‘good’ movie. We first looked at 6.5 stars then at 7 stars to evaluate the accuracy of the perceptron. The perceptron works by computing a weight vector.
If the dot product of the weight vector and the feature vector is greater than some threshold (0 in this case) we classify it as 1, else we classify as 0. The goal of the training phase of the perceptron algorithm is to learn what the weight vector should be to minimize training error (number of misclassified points / total number of points). This algorithm is guaranteed to converge on a set of weights which correctly classify all training points, however there is no time limit for this and as such our algorithm was not allowed to converge.
Secondly we attempted to train a naïve Bayes classifier. This classifier works by using Bayes’s theorem to compute the probability of a particular class given a set of features. This is effectively the same system used in spam detection (P(Spam bunchawords)). The training of this model was the calculation of the probabilities P(Fi C) where Fi is a particular feature (1 or 0) and C is a particular class (1 or 0).
Once both models were trained, we calculated prediction performance for both models. For the perceptron, we simply compute a set of dot products and compare the result according to the above formula. We counted the number of misclassified points and used this to determine error. For Naïve Bayes we computed the formula below for each point:
In the 6.5 trials we tested only for the probability of the movie being good given the actor vector. For 7 we tested both P(Good actors) and P(Bad actors). Our results:
So we're on the right track! We were able to accurately predict whether a movie would be good or bad over 60% of the time.
We believe the error associated with the perceptron (testing error) is not a fully accurate measure of the system being better at 7 than 6.5. Because error is a function of misclassified points, and the probability of misclassification decreases as the boundary between good and bad increases, it becomes easier to predict as the boundary increases. That is, as we increase the boundary from 6.5 to 7, there are many more movies which count as bad than good. This results in it being easier to classify overall without the perceptron needing to be more intelligent.
Moving Forward
In the process of handling and manipulating the data, we had to make several large concessions. First, we had to limit our analysis to ten years to limit the matrix size. It is possible we could get better results looking through fifteen, twenty or even possibly thirty years of data. Secondly, we had to limit the number of features we looked at. Specifically, we only looked at actors within movies, but ignored features like writer, director, budget, time of year, rating and genre. We would like to use these additional metrics to learn better models.
Another large concession we had to make early on was to limit the complexity of our classifier. We had originally intended to attempt to predict actual star ratings rather than just good or bad based on an arbitrary value. This requires a multiclass perceptron which we haven't had time to build. We would also like to add layers to the perceptron, which would improve its ability to learn.
Credits
Much of this post has been snippets and excerpts of our group paper on the subject, and so due credit should be given to Rich and Brian. You can find Brian's site here, and don't forget to check out the continuation of this project on GitHub.
...">Imdb Database Dump Files Oracle(20.03.2020)