Google Will Shut Down Moderator On July 31

Google (fisheye) Google Moderator, a tool for crowdsourcing and ranking user-submitted questions for online and offline events, is shutting down on July 31. The service, which launched in 2008, survived various “spring cleaning” rounds over the years. It was occasionally used by the White House and featured prominently during the 2012 presidential election. Now, however, Google has decided that… Read More

Original URL:

Original article

The Cyberlearning Technologies Transforming Education

aarondubrow writes: The National Science Foundation funds basic cyberlearning research and since 2011 has awarded roughly 170 grants, totaling more than $120 million, to EdTech research projects around the country. However, NSF’s approach to cyber-learning has been different from other public, private and philanthropic efforts. NSF funds compelling ideas, helps rigorously test them and then assists in transitioning the best ideas from research to practice. This article describes several examples of leading cyberlearning projects, from artificial intelligence to augmented reality, that are transforming education.

Share on Google+

Read more of this story at Slashdot.

Original URL:

Original article

Genius Now Lets You Annotate Any Web Page, Not Just Rap

Genius Logo Rap Genius was just the first step to fulfilling Marc Andreessen’s Netscape dream of letting people annotate the whole Internet. Today, the lyric explanation startup tries to achieve its destiny with the launch of its new prefix that lets you add annotations to any web page. Just put the prefix before a URL like , highlight… Read More

Original URL:

Original article

Amazon Elastic File System – Shared File Storage for Amazon EC2

I’d like to give you a preview of a new AWS service that will make its debut later this year.

Let’s take a quick look at the AWS storage portfolio. We currently offer:

Today we are introducing Amazon Elastic File System, our newest storage service. It provides multiple EC2 instances with low-latency, shared access to a fully-managed file system via the popular NFSv4 protocol, and is designed to perform well for a wide variety of workloads, with the ability to scale to thousands of concurrent connections.

We expect to see EFS used for content repositories, development environments, web server farms, home directories, and Big Data applications, to name just a few. If you’ve got the files, we’ve got the storage!

The SSD-based file systems are highly available and highly durable (files, directories, and links are stored redundantly across multiple Availability Zones within an AWS region) and grow or shrink as needed (there’s no need to pre-provision capacity). You’ll be able to create them using the AWS Management Console, the AWS Command Line Interface (CLI), and a simple set of APIs, and start using them within seconds.

File systems can grow to petabyte scale, and throughput and IOPS scale accordingly. You’ll pay only for the storage that you use (billed monthly based on the average used during the month) at the rate of $0.30 per gigabyte per month.

EFS is designed to support the security requirements of large, complex organizations. You’ll be able to use IAM to grant access to the EFS APIs, along with VPC security groups to control network access to file systems. You’ll be able to use standard file and directory permissions (good old chown and chmod) to control access to the directories, files, and links stored on your file systems.

Coming Soon
We will be opening up EFS in preview form in the near future. Visit the Amazon Elastic File System page, and sign up for the preview today, and we will let you know as soon as it is ready for you to use. I will have more information on using Amazon EFS to share with you at that time.


Original URL:

Original article

Amazon Machine Learning – Make Data-Driven Decisions at Scale

Today, it is relatively straightforward and inexpensive to observe and collect vast amounts of operational data about a system, product, or process. Not surprisingly, there can be tremendous amounts of information buried within gigabytes of customer purchase data, web site navigation trails, or responses to email campaigns. The good news is that all of this data can, when properly analyzed, lead to statistically significant results that can be used to make high-quality decisions. The bad news is that you need to find data scientists with relevant expertise in machine learning, hope that your infrastructure is able to support their chosen tool set, and hope (again) that the tool set is sufficiently reliable and scalable for production use.

The science of Machine Learning (often abbreviated as ML) provides the mathematical underpinnings needed to run the analysis and to make sense of the results.  It can help you to turn all of that data into high-quality predictions by finding and codifying patterns and relationships within it. Properly used, Machine Learning can serve as the basis for systems that perform fraud detection (is this transaction legitimate or not?), demand forecasting (how many widgets can we expect to sell?), ad targeting (which ads should be shown to which users?), and so forth.

Introducing Amazon Machine Learning
Today we are introducing Amazon Machine Learning. This new AWS service helps you to use all of that data you’ve been collecting to improve the quality of your decisions. You can build and fine-tune predictive models using large amounts of data, and then use Amazon Machine Learning to make predictions (in batch mode or in real-time) at scale. You can benefit from machine learning even if you don’t have an advanced degree in statistics or the desire to setup, run, and maintain your own processing and storage infrastructure.

I’ll get to the details in just a minute. Before I do so, I’d like to review some of the terminology and the concepts that you need to know in order to fully understand what machine learning does and how you can take advantage of it.

Introduction to Machine Learning
In order to benefit from machine learning, you need to have some existing data that you can use for training purposes. It is helpful to think of the training data as rows in a database or a spreadsheet. Each row represents a single data element (one purchase, one shipment, or one catalog item). The columns represent the attributes of the element:  customer zip code, purchase price, type of credit card, item size, and so forth.

This training data must contain examples of actual outcomes. For example, if you have rows that represent completed transactions that were either legitimate or fraudulent, each row must contain a column that denotes the result, which is also known as the target variable. This data is used to create a Machine Learning Model that, when presented with fresh data on a proposed transaction, will return a prediction regarding its validity. Amazon Machine Learning supports three distinct types of predictions: binary classification, multiclass classification, and regression. Let’s take a look at each one:

Binary classification is used to predict one of two possible outcomes. Is this transaction legitimate, will the customer buy this product, or is the shipping address an apartment complex?

Multiclass classification is used to predict one of three or more possible outcomes and the likelihood of each one. Is this product a book, a movie, or an article of clothing? Is this movie a comedy, a documentary, or a thriller? Which category of products is of most interest to this customer?

Regression is used to predict a number. How many 27″ monitors should we place in inventory? How much should we charge for them? What percent of them are likely to be sold as gifts?

A properly trained and tuned model can be used to answer any one of the questions above. In some cases it is appropriate to use the same training data to build two or more models.

You should plan to spend some time enriching your data in order to ensure that it is a good match for the training process. As a simple example, you might start out with location data that is based on zip or postal codes. After some analysis, you could very well discover that you can improve the quality of the results by using a different location representation that contains greater or lesser resolution. The ML training process is iterative; you should definitely plan to spend some time understanding and evaluating your initial results and then using them to enrich your data.

You can measure the quality of each of your models using a set of performance metrics that are computed and made available to you. For example, the Area Under Curve (AUC) metric monitors the performance of a binary classification. This is a floating point value in the range 0.0 to 1.0 that expresses how often the model predicts the correct answer on data it was not trained with. Values rise from 0.5 to 1.0 as as the quality of the model rises. A score of 0.5 is no better than random guessing, while 0.9 would be a very good model for most cases. But a score of 0.9999 is probably too good to be true, and might indicate a problem with the training data.

As you build your binary prediction model, you will need to spend some time looking at the results and adjusting a value known as the cut-off. This represents the probability that the prediction is true; you can adjust it up or down based on the relative importance of false positives (predictions that should be false but were predicted as true) and false negatives (predictions that should be true but were predicted as false) in your particular situation. If you are building a spam filter for email, a false negative will route a piece of spam to your inbox and a false positive will route legitimate mail to your junk folder. In this case, false positives are undesirable. The tradeoffs between false positives and false negatives is going to depend on your business problem and how you plan to make use of the model in a production setting,

Amazon Machine Learning in Action
Let’s walk through the process of creating a model and generating some predictions using the steps described in the Tutorial section of the Amazon Machine Learning Developer Guide. You can sign up for Amazon Machine Learning and follow the steps in the guide on your own if you would like. The guide uses a lightly enhanced copy of the publicly available bank marketing dataset from the UC Irvine Machine Learning Repository. The model that we will build will answer the question “Will the customer subscribe to my new product?”

I downloaded a copy of  banking.csv  and uploaded it to Amazon Simple Storage Service (S3), and then agreed to allow the console to add an IAM policy so that Amazon Machine Learning could access it:

Then I created an Amazon Machine Learning Datasource object by referencing the item in the bucket and supplying a name for the object. This object holds the location of the data, the variable names and types, the name of the target variable, and descriptive statistics for each variable. Most of the operations within Amazon Machine Learning reference a Datasource. Here’s how I set everything up:

Amazon Machine Learning can also create a Datasource from Amazon Redshift or an Amazon RDS MySQL database. Selecting the Redshift option shown above would have given me the option to enter the name of my Amazon Redshift cluster, along with a database name, access credentials, and a SQL query.   The Machine Learning API can be used to create a Datasource from an Amazon RDS for MySQL database.

Amazon Machine Learning opened and scanned the file, made a guess at the variable types, and then proposed the following schema:

In this case all of its guesses were correct. If they weren’t, I could have selected one or more rows and clicked on Change Type to fix them.

Because I will use the Datasource to create and evaluate an ML model, I need to select the training variable. In this data set, the training variable (y) has the Binary data type so the models generated from it will use binary classification.

After a couple more clicks I was ready to create my Datasource:

My Datasource was ready a minute or two later:

As I hinted at earlier, you can improve your models by knowing more about your data. The Amazon Machine Learning Console provides you with several different tools that you can use to learn more. For example, you can look at the distribution of values for any of the variables in a Datasource. Here’s what I saw when I inspected the age variable in mine:

The next step is to create my model:

I chose to use the default settings. Amazon Machine Learning used 70% of the data for training and 30% to evaluate the model based on the defaults:

If I had chosen the Custom option, I would have had the opportunity to customize the “recipe” that Amazon Machine Learning uses to transform and process the data from the Datasource:

After a couple more clicks, Amazon Machine Learning started to create my model. I took a quick break to water my vegetable garden and returned to find that my model was ready to go:

I took a quick look at the performance metrics:

In order to select the best customers (those most likely to make a purchase), I clicked on Adjust Score Threshold and bumped up the cut-off value until 5% of the records were expected to pass by predicting a value of “1” for y:

With this setting, only 1.3% of the predictions will be false positives, 22% will be false negatives, and 77% will be correct predictions.  I chose to treat false positives as expensive, and decided to set the cut off value high in order to avoid them. In business terms this setting allowed me to avoid sending costly promotional materials to the “wrong” customers.

With my model built, I can now use it to create batch predictions (recall that Amazon Machine Learning supports both batch and real-time predictions). The batch model allows me to generate a set of predictions for a set of observations all at once. I start from the menu:

I created another Datasource using the file recommended by the Getting Started Guide. This file, unlike the first one, does not contain values for the y variable.

Then I selected a location (in S3) to store the predictions, reviewed my choices, and initiated the batch prediction:

After another quick trip to the garden, my predictions were ready! I downloaded the file from the bucket, unzipped it, and here’s what I saw:

The rows correspond to those in the original file. The first value is the predicted y variable (computed by comparing the prediction score against the cut off that I set when I was building the model), and the second is the actual score. If I had included a Row Identifier, each prediction would include a unique “primary key” that would allow me to tie the results back to the source data.

If I am building a real time application and I need to generate predictions as part of a request-response cycle, I can enable the model for real time predictions like this:

After real time predictions have been enable, I can write code to call Amazon Machine Learning‘s Predict function. Here’s some Java code that retrieves the metadata associated with an ML model (mlModelId in the code), finds the service endpoint in the metadata, makes a real-time prediction, and displays the result:

AmazonMachineLearningClient client = new AmazonMachineLearningClient();
GetMLModelRequest modelRequest = new GetMLModelRequest()
GetMLModelResult model = client.getMLModel(modelRequest);
String predictEndpoint = model.getEndpointInfo().getEndpointUrl();
PredictRequest predictRequest = new PredictRequest()
 PredictResult prediction = client.predict(predictRequest);

The code will produce output that looks like this:

{Prediction: {PredictedLabel: 0,PredictedScores: {0=0.10312237},Details: {PredictiveModelType=BINARY, Algorithm=SGD}}}

This means that the ML model type was Binary classification, the predicted score was 0.10312237, and based on the prediction threshold associated with the model when the code was run, the predicted response was ‘0’.

Check out our Machine Learning Samples to see some sample code (Python and Java) for targeted marketing, social media monitoring, and mobile predictions.

Things to Know
Amazon Machine Learning is available now and you can start using it today in the US East (Northern Virginia) region.

Pricing, as usual, is on a pay-as-you-go basis:

  • Data analysis, model training, and model evaluation will cost you $0.42 per compute hour.
  • Batch predictions will cost $0.10 for every 1,000 predictions, rounded up to the next 1,000.
  • Real time predictions cost $0.10 for every 1,000 predictions  plus an hourly reserved capacity charge of $0.001 per hour for each 10 MB of memory provisioned for your model. During model creation, you specify the maximum memory size of each model to manage the cost and to control predictive performance.
  • Charges for data stored in S3, Amazon RDS, and Amazon Redshift are billed separately.

Learning More
To learn more about Amazon Machine Learning, read the [Amazon Machine Learning] documentation!

— Jeff;

Original URL:

Original article

LinkedIn buys training site, enters professional development market

LinkedIn is acquiring online learning company for $1.5 billion in cash and stock, the social networking site announced Thursday. offers professional development courses on design, creative and business topics. Some of the courses, for example, teach how to write HTML, negotiate better, or use design software like Photoshop. The site was launched by Lynda Weinman and Bruce Heavin in 1995 as a way to teach Web publishing and design.

Integrating with LinkedIn would allow job seekers to know what skills are required for a position they’re interested in and immediately be prompted to take a course in that subject, Ryan Roslansky, LinkedIn’s head of content, said in a blog post. Further details on whether would be combined with LinkedIn, or if the training site would continue to operate independently weren’t provided.

To read this article in full or to leave a comment, please click here

Original URL:

Original article

Amazon offers network file storage in the cloud

Amazon Web Services continues to chip away at the enterprise storage market, with plans for a new service designed to be a nimbler alternative to network attached storage (NAS) appliances.

The Amazon Elastic File System (EFS) will provide a shared, low-latency file system for project teams and organizations, such as a video production company, that need to share large files and access them quickly.

“The file system is the missing element in the cloud today,” Amazon Web Services head Andy Jassy said Thursday at the AWS Summit in San Francisco. The service is not yet available for full commercial use, though a preview will be available soon.

To read this article in full or to leave a comment, please click here

Original URL:

Original article

AWS offers machine-learning service to make sense of big data

Enterprises will soon be able to do the kind of big data analysis that enables Amazon to recommend books, video games and toasters to its customers.

Amazon Web Services (AWS) announced Thursday at its cloud summit in San Francisco that it’s rolling out Amazon Machine Learning, a fully managed, cloud-based service designed to pull useful information from mountains of data.

The problem with big data is that it often simply sits there unused because it’s far too complicated and energy- and time-intensive to find the critical information hidden inside.

AWS, following in the footsteps of cloud competitor Microsoft wants its new cloud service to help with that. Microsoft added a machine learning service to Azure in February.

To read this article in full or to leave a comment, please click here

Original URL:

Original article

Proudly powered by WordPress | Theme: Baskerville 2 by Anders Noren.

Up ↑

%d bloggers like this: