学习笔记(四): Representation:Feature Engineering/Qualities of Good Features/Cleaning Data/Feature Sets
目录
Avoid rarely used discrete feature values
Prefer clear and obvious meanings
Don't mix "magic" values with actual data
Account for upstream instability上游不稳定性
Learn more about binning boundaries
Representation
A machine learning model can't directly see, hear, or sense input examples. Instead, you must create a representation of the data to provide the model with a useful vantage point into the data's key qualities. That is, in order to train a model, you must choose the set of features that best represent the data.
Learning Objectives
- Map fields from logs and protocol buffers into useful ML features.
- Determine which qualities comprise great features.
- Handle outlier features.
- Investigate the statistical properties of a data set.
- Train and evaluate a model with tf.estimator.
Feature Engineering
In traditional programming, the focus is on code. In machine learning projects, the focus shifts to representation.
That is, one way developers hone a model is by adding and improving its features.
Mapping Raw Data to Features
The left side of Figure 1 illustrates raw data from an input data source; the right side illustrates a feature vector, which is the set of floating-point values comprising the examples in your data set.
Feature engineering means transforming raw data into a feature vector. Expect to spend significant time doing feature engineering.
Many machine learning models must represent the features as real-numbered vectors since the feature values must be multiplied by the model weights.
Figure 1. Feature engineering maps raw data to ML features.
Mapping numeric values
Integer and floating-point data don't need a special encoding because they can be multiplied by a numeric weight. As suggested in Figure 2, converting the raw integer value 6 to the feature value 6.0 is trivial:
Figure 2. Mapping integer values to floating-point values.
Mapping categorical values
Categorical features have a discrete set of possible values. For example, there might be a feature called street_name
with options that include:
{'Charleston Road', 'North Shoreline Boulevard', 'Shorebird Way', 'Rengstorff Avenue'}
Since models cannot multiply strings by the learned weights, we use feature engineering to convert strings to numeric values.
We can accomplish this by defining a mapping from the feature values, which we'll refer to as the vocabulary of possible values, to integers.
Since not every street in the world will appear in our dataset, we can group all other streets into a catch-all "other" category, known as an OOV (out-of-vocabulary) bucket.
Using this approach, here's how we can map our street names to numbers:
- map Charleston Road to 0
- map North Shoreline Boulevard to 1
- map Shorebird Way to 2
- map Rengstorff Avenue to 3
- map everything else (OOV) to 4
However, if we incorporate these index numbers directly into our model, it will impose some constraints that might be problematic:
We'll be learning a single weight that applies to all streets. For example, if we learn a weight of 6 for
street_name
, then we will multiply it by 0 for Charleston Road, by 1 for North Shoreline Boulevard, 2 for Shorebird Way and so on. Consider a model that predicts house prices usingstreet_name
as a feature. It is unlikely that there is a linear adjustment of price based on the street name, and furthermore this would assume you have ordered the streets based on their average house price. Our model needs the flexibility of learning different weights for each street that will be added to the price estimated using the other features.We aren't accounting for cases where
street_name
may take multiple values. For example, many houses are located at the corner of two streets, and there's no way to encode that information in thestreet_name
value if it contains a single index.
To remove both these constraints, we can instead create a binary vector for each categorical feature in our model that represents values as follows:
- For values that apply to the example, set corresponding vector elements to
1
. - Set all other elements to
0
.
The length of this vector is equal to the number of elements in the vocabulary. This representation is called a one-hot encoding when a single value is 1, and a multi-hot encoding when multiple values are 1.
Figure 3 illustrates a one-hot encoding of a particular street: Shorebird Way.
The element in the binary vector for Shorebird Way has a value of 1
, while the elements for all other streets have values of 0
.
Figure 3. Mapping street address via one-hot encoding.
This approach effectively creates a Boolean variable for every feature value (e.g., street name). Here, if a house is on Shorebird Way then the binary value is 1 only for Shorebird Way. Thus, the model uses only the weight for Shorebird Way.
Similarly, if a house is at the corner of two streets, then two binary values are set to 1, and the model uses both their respective weights.
映射分类(枚举)值
分类特征具有一组离散的可能值。例如,名为
Lowland Countries
的特征只包含 3 个可能值:{'Netherlands', 'Belgium', 'Luxembourg'}您可能会将分类特征(如 Lowland Countries)编码为枚举类型或表示不同值的整数离散集。例如:将荷兰表示为 0,将比利时表示为 1,将卢森堡表示为 2。
不过,机器学习模型通常将每个分类特征表示为单独的布尔值。例如,Lowland Countries 在模型中可以表示为 3 个单独的布尔值特征:x1:是荷兰吗?x2:是比利时吗?x3:是卢森堡吗?
采用这种方法编码还可以简化某个值可能属于多个分类这种情况(例如,“与法国接壤”对于比利时和卢森堡来说都是 True)。
One-hot encoding extends to numeric data that you do not want to directly multiply by a weight, such as a postal code.
Sparse Representation
Suppose that you had 1,000,000 different street names in your data set that you wanted to include as values for street_name
. Explicitly creating a binary vector of 1,000,000 elements where only 1 or 2 elements are true is a very inefficient representation in terms of both storage and computation time when processing these vectors.
In this situation, a common approach is to use a sparse representation in which only nonzero values are stored. In sparse representations, an independent model weight is still learned for each feature value, as described above.
Glossay
1.discrete feature:A feature with a finite set of possible values. For example, a feature whose values may only be animal, vegetable, or mineral is a discrete (or categorical) feature. Contrast with continuous feature.
2.feature engineering:
The process of determining which features might be useful in training a model确定哪些特征可能在训练模型方面非常有用, and then converting raw data from log files and other sources into said features.然后将日志文件及其他来源的原始数据转换为所需的特征。In TensorFlow, feature engineering often means converting raw log file entries原始日志文件条目 to tf.Example protocol buffers.See also tf.Transform.Feature engineering is sometimes called feature extraction.
3.one-hot encoding:A sparse vector稀疏向量 in which:One element is set to 1.All other elements are set to 0.
One-hot encoding is commonly used to represent strings or identifiers字符串或标识符 that have a finite set of possible values.
For example, suppose a given botany data set chronicles 15,000 different species, each denoted with a unique string identifier. As part of feature engineering, you'll probably encode those string identifiers as one-hot vectors in which the vector has a size of 15,000.
4.representation:The process of mapping data to useful features.
5.sparse representation:A representation of a tensor that only stores nonzero elements.For example, the English language consists of about a million words. Consider two ways to represent a count of the words used in one English sentence:
- A dense representation of this sentence must set an integer for all one million cells, placing a 0 in most of them, and a low integer into a few of them.
- A sparse representation of this sentence stores only those cells symbolizing a word actually in the sentence. So, if the sentence contained only 20 unique words, then the sparse representation for the sentence would store an integer in only 20 cells.
For example, consider two ways to represent the sentence, "Dogs wag tails." As the following tables show, the dense representation consumes about a million cells; the sparse representation consumes only 3 cells:
Dense Representation Cell Number Word Occurrence 0 a 0 1 aardvark 0 2 aargh 0 3 aarti 0 … 140,391 more words with an occurrence of 0 140395 dogs 1 … 633,062 words with an occurrence of 0 773458 tails 1 … 189,136 words with an occurrence of 0 962594 wag 1 … many more words with an occurrence of 0
Sparse Representation Cell Number Word Occurrence 140395 dogs 1 773458 tails 1 962594 wag 1
Qualities of Good Features
We've explored ways to map raw data into suitable feature vectors, but that's only part of the work. We must now explore what kinds of values actually make good features within those feature vectors.
Avoid rarely used discrete feature values
Good feature values should appear more than 5 or so times in a data set. Doing so enables a model to learn how this feature value relates to the label. That is, having many examples with the same discrete value gives the model a chance to see the feature in different settings, and in turn, determine when it's a good predictor for the label. For example, a house_type
feature would likely contain many examples in which its value was victorian
:
✔This is a good example:
house_type: victorian
Conversely, if a feature's value appears only once or very rarely, the model can't make predictions based on that feature. For example, unique_house_id
is a bad feature because each value would be used only once, so the model couldn't learn anything from it:
✘The following is an example of a unique value. This should be avoided.
unique_house_id: 8SK982ZZ1242Z
Prefer clear and obvious meanings
Each feature should have a clear and obvious meaning to anyone on the project. For example, consider the following good feature for a house's age, which is instantly recognizable as the age in years:
✔The following is a good example of a clear value.
house_age: 27
Conversely, the meaning of the following feature value is pretty much indecipherable难以理解的 to anyone but the engineer who created it:
✘The following is an example of a value that is unclear. This should be avoided。
house_age: 851472000
In some cases, noisy data (rather than bad engineering choices) causes unclear values. For example, the following user_age came from a source that didn't check for appropriate values:
✘The following is an example of noisy/bad data. This should be avoided.
user_age: 277
Don't mix "magic" values with actual data
Good floating-point features don't contain peculiar奇特 out-of-range discontinuities or "magic" values. For example, suppose a feature holds a floating-point value between 0 and 1. So, values like the following are fine:
✔The following is a good example:
quality_rating: 0.82
quality_rating: 0.37
However, if a user didn't enter a quality_rating
, perhaps the data set represented its absence with a magic value like the following:
✘The following is an example of a magic value. This should be avoided.
quality_rating: -1
To work around magic values, convert the feature into two features:
- One feature holds only quality ratings, never magic values.
- One feature holds a boolean value indicating whether or not a
quality_rating
was supplied. Give this boolean feature a name likeis_quality_rating_defined
.
Account for upstream instability上游不稳定性
The definition of a feature shouldn't change over time. For example, the following value is useful because the city name probably won't change. (Note that we'll still need to convert a string like "br/sao_paulo" to a one-hot vector.)
✔This is a good example:
city_id: "br/sao_paulo"
But gathering a value inferred by another model carries additional costs. Perhaps the value "219" currently represents Sao Paulo, but that representation could easily change on a future run of the other model:
✘The following is an example of a value that could change. This should be avoided.
inferred_city_cluster: "219"
Cleaning Data
Apple trees produce some mixture of great fruit and wormy messes. Yet the apples in high-end grocery stores display 100% perfect fruit. Between orchard果园 and grocery, someone spends significant time removing the bad apples or throwing a little wax on the salvageable挽救 ones. As an ML engineer, you'll spend enormous amounts of your time tossing out折腾 bad examples and cleaning up the salvageable ones. Even a few "bad apples" can spoil a large data set.
Scaling feature values
Scaling缩放 means converting floating-point feature values from their natural range (for example, 100 to 900) into a standard range (for example, 0 to 1 or -1 to +1). If a feature set consists of only a single feature, then scaling provides little to no practical benefit. If, however, a feature set consists of multiple features, then feature scaling provides the following benefits:
- Helps gradient descent converge汇合 more quickly.
- Helps avoid the "NaN trap," in which one number in the model becomes a NaN (e.g., when a value exceeds the floating-point precision limit during training), and—due to math operations—every other number in the model also eventually becomes a NaN.
- Helps the model learn appropriate weights for each feature. Without feature scaling, the model will pay too much attention to the features having a wider range.
You don't have to give every floating-point feature exactly the same scale. Nothing terrible will happen if Feature A is scaled from -1 to +1 while Feature B is scaled from -3 to +3. However, your model will react poorly if Feature B is scaled from 5000 to 100000.
Learn more about scaling
One obvious way to scale numerical data is to linearly map [min value, max value] to a small scale, such as [-1, +1].
Another popular scaling tactic is to calculate the Z score of each value. The Z score relates the number of standard deviations away from the mean. In other words:
scaledvalue=(value−mean)/stddev.
For example, given:
- mean = 100
- standard deviation = 20
- original value = 130
then:
scaled_value = (130 - 100) / 20
scaled_value = 1.5Scaling with Z scores means that most scaled values will be between -3 and +3, but a few values will be a little higher or lower than that range.
Handling extreme outliers
The following plot represents a feature called roomsPerPerson
from the California Housing data set. The value of roomsPerPerson
was calculated by dividing the total number of rooms for an area by the population for that area. The plot shows that the vast majority of areas in California have one or two rooms per person. But take a look along the x-axis.
Figure 4. A verrrrry lonnnnnnng tail.
How could we minimize the influence of those extreme outliers? Well, one way would be to take the log of every value:
Figure 5. Logarithmic scaling对数缩放 still leaves a tail.
Log scaling does a slightly better job, but there's still a significant tail of outlier values. Let's pick yet another approach. What if we simply "cap" or "clip" the maximum value of roomsPerPerson
at an arbitrary value, say 4.0?
Figure 6. Clipping feature values at 4.0
Clipping the feature value at 4.0 doesn't mean that we ignore all values greater than 4.0. Rather, it means that all values that were greater than 4.0 now become 4.0. This explains the funny hill at 4.0. Despite that hill, the scaled feature set is now more useful than the original data.
Binning分档
The following plot shows the relative prevalence流行 of houses at different latitudes in California. Notice the clustering—Los Angeles is about at latitude 34 and San Francisco is roughly at latitude 38.
Figure 7. Houses per latitude.
In the data set, latitude
is a floating-point value. However, it doesn't make sense to represent latitude
as a floating-point feature in our model. That's because no linear relationship exists between latitude and housing values. For example, houses in latitude 35 are not 35/34 more expensive (or less expensive) than houses at latitude 34. And yet, individual latitudes probably are a pretty good predictor of house values.
To make latitude a helpful predictor, let's divide latitudes into "bins" as suggested by the following figure:
Figure 8. Binning values.
Instead of having one floating-point feature, we now have 11 distinct boolean features (LatitudeBin1
, LatitudeBin2
, ..., LatitudeBin11
). Having 11 separate features is somewhat inelegant, so let's unite them into a single 11-element vector. Doing so will enable us to represent latitude 37.4 as follows:[0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0]
Thanks to binning, our model can now learn completely different weights for each latitude.
Learn more about binning boundaries
For simplicity's sake in the latitude example, we used whole numbers as bin boundaries. Had we wanted finer-grain resolution, we could have split bin boundaries at, say, every tenth of十分之一 a degree. Adding more bins enables the model to learn different behaviors from latitude 37.4 than latitude 37.5, but only if there are sufficient examples at each tenth of a latitude.
Another approach is to bin by quantile分位数, which ensures that the number of examples in each bucket is equal. Binning by quantile completely removes the need to worry about outliers.
Scrubbing
Until now, we've assumed that all the data used for training and testing was trustworthy. In real-life, many examples in data sets are unreliable due to one or more of the following:
- Omitted values. For instance, a person forgot to enter a value for a house's age.
- Duplicate examples. For example, a server mistakenly uploaded the same logs twice.
- Bad labels. For instance, a person mislabeled a picture of an oak tree as a maple.
- Bad feature values. For example, someone typed in an extra digit, or a thermometer was left out in the sun.
Once detected, you typically "fix" bad examples by removing them from the data set. To detect omitted values or duplicated examples, you can write a simple program. Detecting bad feature values or labels can be far trickier.
In addition to detecting bad individual examples, you must also detect bad data in the aggregate. Histograms are a great mechanism机制 for visualizing your data in the aggregate. In addition, getting statistics like the following can help:
- Maximum and minimum
- Mean and median
- Standard deviation
Consider generating lists of the most common values for discrete features. For example, do the number of examples with country:uk
match the number you expect. Should language:jp
really be the most common language in your data set?
Know your data
Follow these rules:
- Keep in mind what you think your data should look like.
- Verify that the data meets these expectations (or that you can explain why it doesn’t).
- Double-check that the training data agrees with other sources (for example, dashboards).
Treat your data with all the care that you would treat any mission-critical code. Good ML relies on good data.
Additional Information
Rules of Machine Learning, ML Phase II: Feature Engineering
Glossay
- binning:See bucketing.
- bucketing:Converting a (usually continuous) feature into multiple binary features called buckets or bins, typically based on value range. For example, instead of representing temperature as a single continuous floating-point feature, you could chop ranges of temperatures into discrete bins. Given temperature data sensitive to a tenth of a degree, all temperatures between 0.0 and 15.0 degrees could be put into one bin, 15.1 to 30.0 degrees could be a second bin, and 30.1 to 50.0 degrees could be a third bin.
- NaN trap:When one number in your model becomes a NaN during training, which causes many or all other numbers in your model to eventually become a NaN.NaN is an abbreviation for "Not a Number."
- scaling:A commonly used practice in feature engineering to tame a feature's range of values to match the range of other features in the data set. For example, suppose that you want all floating-point features in the data set to have a range of 0 to 1. Given a particular feature's range of 0 to 500, you could scale that feature by dividing each value by 500.See also normalization.
- normalization:The process of converting an actual range of values into a standard range of values, typically -1 to +1 or 0 to 1. For example, suppose the natural range of a certain feature is 800 to 6,000. Through subtraction and division, you can normalize those values into the range -1 to +1.See also scaling.
- feature set:The group of features your machine learning model trains on. For example, postal code, property size, and property condition might comprise a simple feature set for a model that predicts housing prices.
- outliers:Values distant from most other values. In machine learning, any of the following are outliers:1.Weights with high absolute values高绝对值.2.Predicted values relatively far away from the actual values.3.Input data whose values are more than roughly 3 standard deviations 三个标准差from the mean. Outliers often cause problems in model training.
Programming Exercise
In this programming exercise, you'll create a good, minimal set of features.
学习笔记(四): Representation:Feature Engineering/Qualities of Good Features/Cleaning Data/Feature Sets的更多相关文章
- C#可扩展编程之MEF学习笔记(四):见证奇迹的时刻
前面三篇讲了MEF的基础和基本到导入导出方法,下面就是见证MEF真正魅力所在的时刻.如果没有看过前面的文章,请到我的博客首页查看. 前面我们都是在一个项目中写了一个类来测试的,但实际开发中,我们往往要 ...
- IOS学习笔记(四)之UITextField和UITextView控件学习
IOS学习笔记(四)之UITextField和UITextView控件学习(博客地址:http://blog.csdn.net/developer_jiangqq) Author:hmjiangqq ...
- java之jvm学习笔记四(安全管理器)
java之jvm学习笔记四(安全管理器) 前面已经简述了java的安全模型的两个组成部分(类装载器,class文件校验器),接下来学习的是java安全模型的另外一个重要组成部分安全管理器. 安全管理器 ...
- Learning ROS for Robotics Programming Second Edition学习笔记(四) indigo devices
中文译著已经出版,详情请参考:http://blog.csdn.net/ZhangRelay/article/category/6506865 Learning ROS for Robotics Pr ...
- Typescript 学习笔记四:回忆ES5 中的类
中文网:https://www.tslang.cn/ 官网:http://www.typescriptlang.org/ 目录: Typescript 学习笔记一:介绍.安装.编译 Typescrip ...
- ES6学习笔记<四> default、rest、Multi-line Strings
default 参数默认值 在实际开发 有时需要给一些参数默认值. 在ES6之前一般都这么处理参数默认值 function add(val_1,val_2){ val_1 = val_1 || 10; ...
- muduo网络库学习笔记(四) 通过eventfd实现的事件通知机制
目录 muduo网络库学习笔记(四) 通过eventfd实现的事件通知机制 eventfd的使用 eventfd系统函数 使用示例 EventLoop对eventfd的封装 工作时序 runInLoo ...
- python3.4学习笔记(四) 3.x和2.x的区别,持续更新
python3.4学习笔记(四) 3.x和2.x的区别 在2.x中:print html,3.x中必须改成:print(html) import urllib2ImportError: No modu ...
- Go语言学习笔记四: 运算符
Go语言学习笔记四: 运算符 这章知识好无聊呀,本来想跨过去,但没准有初学者要学,还是写写吧. 运算符种类 与你预期的一样,Go的特点就是啥都有,爱用哪个用哪个,所以市面上的运算符基本都有. 算术运算 ...
随机推荐
- linux 之基本命令学习总结
前言:从今天开始写这系列linux博客了(是学习刘遄老师的<linux就该这么学>),视频学习的资源可以在b站上找到:https://www.bilibili.com/video/av45 ...
- java基础第五篇封装与面向对象
a.方法: public static void main(String[] args) { } 一般定义标准: 形参:一般把 不确定的量或者变化的量定义在形参位置//圆的的半径,长方形的长和宽,传递 ...
- glassfish 自定义 jaas realm
https://www.oschina.net/translate/custom-jaas-realm-for-glassfish-3
- Maven项目已启动但是报异常访问webapp下所有资源都404
- Maven项目编译版本的问题和Spring中没有导入核心包
idea中maven项目的编译: 方案1:maven的settings.xml中指定全局默认编译版本 <profile> <id>jdk-1.8</id> < ...
- Spring创建对象的几种方法
一.通过构造器 无参构造器 <?xml version="1.0" encoding="UTF-8"?> <beans xmlns=" ...
- python之简单登陆
简单的登陆 要求: 1.用户名输错3次,系统退出 2.密码输错3次,账号被禁用 3.使用文件操作知识 流程图: 项目代码: def login(): i=0 while i <3 : name ...
- 机器学习框架ML.NET学习笔记【9】自动学习
一.概述 本篇我们首先通过回归算法实现一个葡萄酒品质预测的程序,然后通过AutoML的方法再重新实现,通过对比两种实现方式来学习AutoML的应用. 首先数据集来自于竞赛网站kaggle.com的UC ...
- 路径方案数(mod)
路径方案数(mod) [题目描述] 给一张无向图,n 个点和 m 条边,cyb 在 1 号点,他要去 2 号点, cyb 可以从 a 走到 b,当且仅当a到2的最短路,比b 到2的最短路长. 求 cy ...
- FastDFS整合Nginx
浏览器访问FastDFS存储的图片需要通过Nginx访问 需要安装fastdfs-nginx-module.Nginx 安装fastdfs-nginx-module模块 (1)将fastdfs-ngi ...