FAQ
===
.. raw:: html
.. faq-topic:: doubletaxonomy
.. pretty-faq::
:header: * How does PayScale maintain the proprietary taxonomy?
We maintain our proprietary taxonomy through a series of steps summarized below. These steps ensure usable data
that is both accurate and granular, which allows us to make pay predictions.
To normalize the crowdsourced data that we collect from our online salary survey, we maintain proprietary group
classifications based on repudiated third-party structures that are industry standard for the United States.
We have an automated rules engine that applies this taxonomy to the factors in our database that impact pay.
This vastly improves the accuracy of our salary predications because similarity between individual classes of
answers is baked-in. We save the most specific answer, but once the taxonomy has been applied we can easily
move up and down the hierarchy for things like job titles and locations when data is sparse. We maintain
internal classifications for Job Titles (groupings loosely related to ONET codes), Industries (groupings
related to NAICS), Locations (city metro definitions), Education, Employer Names, Skills, and Certifications.
These answers are maintained using a combination of automation and human validation. We review our taxonomy
once a quarter to check for new answers and to review client feedback in order to make updates. Taxonomy
updates, depending on the nature, can take time to process because all of our algorithms must be updated to
reflect the changes before they are available via API or in product. When adding new answers, changes will
be reflected when we capture enough data via our survey to model the results.
|
.. pretty-faq::
:header: * Why does the taxonomy matter?
On a basic level, applying a taxonomy makes the data usable. If we didn't apply the taxonomy we would just have a
bunch of disparate strings that were similar (Analyst, Data - Data Analyst - Data Analyst II) but
maybe not considered to be the same entity.
|
.. faq-topic:: light-data
.. pretty-faq::
:header: * How is PayScale able to predict pay in places where you have light or no data?
PayScale has the ability to pull in similar entities when data is somewhat
low for the given query. In some platforms, data might be limited to a specific
city or a specific job title, but with our taxonomy, we generally know what
cities fall in which metros and can pull in more data at a granular level to make
a prediction. Job titles, in particular, might not exist in sufficient numbers
in certain geographic locations, but we know how similar jobs are paid in that area.
|
.. faq-topic:: _data-coverage
.. pretty-faq::
:header: * What is the country/geographic coverage of the queries?
PayScale only has a compensation model built for 8 countries:
US, Canada, Australia, New Zealand, South Africa, UK, Ireland and India.
Our data is strongest in English-speaking countries (or in India’s case,
where English is not the primary language, however hosts 125 million
English-speakers) due to our survey being in English. Therefore, data we
collect from non-English speaking countries have a tendency to not be an
accurate representation of the labor force (i.e., biased), as it is
from those who have either learned English or ex-pats residing and
working in a foreign country.
|
.. faq-topic:: _data-maintenance
.. pretty-faq::
:header: * How you maintain and update the data and taxonomy/ontologies that power the API?
In terms of our promise of fresh data, we include relevant profiles in a
market report collected as recently as the previous day. Our model does
place a higher emphasis on more recent profiles when trying to find the
best matched 45 profiles, but will extend the time horizon to ensure we
get the most relevant profiles to the position described, while never going
back more than 2 years. For more information on why 45 profiles are used,
see :faq-topic:`Why 45 Profiles? `.
|
.. faq-topic:: input-influence
.. pretty-faq::
:header: * How does a submitted Answer influence the Reports I receive from the API?
Answers submitted in the Request are used by the Compensation Model to
select the best 45 Profiles which are used to generate the pay distributions
that make up the values in the Reports. While not every Answer is used
by every Report (for instance, a multi-request to ["pay", "yoe"] with
a YearsExperience value of 5 will influence the PayReport, but not the
YearsExperienceReport), the majority are cross-compatible.
|
.. report-question:
.. pretty-faq::
:header: * What does from my Report mean?
For a detailed description of every response in every Report, please check
out the :ref:`Reports Page `.
|
.. faq-topic:: what-is-job-title-rating
.. pretty-faq::
:header: * What is JobTitleRating?
In short, the JobTitleRating is a score from 0 to 100 that represents
the confidence our algorithm has that the submitted JobTitle is accurately
described by the MatchedJobTitle for which the Report was run.
For a detailed description of every response in every Report, please check
out the :ref:`jobtitle-rating` page.
|
.. faq-topic:: _base-pay-report
.. pretty-faq::
:header: * What is "Base Pay" in a Base Pay Report?
Base Pay, or Salary, is defined as the salary reported by a set of Profiles
from our crowdsourced data. In general, Salary is defined as the recurring
annual pay received by a worker without additional Bonuses or Incentives.
|
.. faq-topic:: profiles
.. pretty-faq::
:header: * Why 45 Profiles?
The 45 most relevant Profiles offer a nice balance between statistical
power for accurate reporting, coverage when data is sparse, and signal
for a particular sub-set of the data. In the limit, if only one Profile
was required, getting a distribution would be impossible. If, say, two
hundred Profiles were required, sparse data would be overpowered by similar,
but less relevant, data. By choosing 45, a more intermediate value which was
calibrated during the creation of the Compensation Model, a good balance is
achieved between signal, noise, and coverage.
|
.. faq-topic:: auto-resolve-title
.. pretty-faq::
:header: * How Do I Know What My Job Title Auto-Resolved To?
Any Report which has the optional argument AutoResolveJobTitle will return
a Context object in the response if AutoResolveJobTitle is True. This Context
object, described on the :ref:`Reports Page `, contains
a SubmittedJobTitle (the JobTitle submitted in the request, which may or may
not be a PayScale Confirmed JobTitle), and MatchedJobTitle (a PayScale
Confirmed JobTitle which is our best guess at mapping the SubmittedJobTitle to
our internal taxonomy).
In the case where this guess is poor, human intervention may be required
in order to search for a more fitting JobTitle to submit. There are two endpoints
that can help with this - :ref:`JobTitleMatch `, which hits the
same service as AutoResolveJobTitle, and :ref:`AutoComplete `,
which exposes a more direct, syntactic search method.
|
.. faq-topic:: negative-differential
.. pretty-faq::
:header: * Why Can A Skill Differential Be Negative?
Skill differentials, or Skill Impacts, are calculated against the median
Base Pay from a cohort of Profiles associated with the given JobTitle. Because
they are calculated against the median of the entire cohort, and because Skills are not randomly
distributed, there will naturally be Skills associated with positive impact
and there will be Skills associated with a negative impact, where the median
Base Pay associated with Profiles for which that Skill is present is below
the median Base Pay for all Profiles in the cohort.
When this lower median associated with a Skill is compared to the higher median
associated with the entire cohort, the resulting differential will be negative.
Other differentials (including Certifications) are calculated the same way.
|