Category: Statistics

  • Streaming services spending analysis

    Instructions REMEMBER: You will be submitting 2 files – an Excel file with your work AND a Word file with your report. You will receive a zero (0) for the assignment if you don’t submit both parts! Your report should contain the Excel tables, graphics, and statistics you are asked to create by using copy/paste functionality from Excel to Word. You can also use “snipping tool” as an alternative for copying from Excel to Word. Use the Excel file M07 Business Application – Streaming ServicesDownload M07 Business Application – Streaming Services Part 1 A random sample of 25 customers of streaming services was surveyed and asked how much they spend on streaming services monthly. Other information gathered is the customer age, monthly household income, and family size. Your job is to do an analysis to find the best way to estimate the amount spent on streaming services based on the data you have. Here are your instructions to guide your work: Create a correlation matrix of all the variables from the file. Which variable has the strongest linear correlation with the amount spent on streaming services? Explain this in your report. Create a simple linear regression equation which can estimate the amount spent (this is the y variable) and your chosen variable from the previous step (this is the x variable). State your equation and include Excels Analysis Toolpak Regression output in your report. Interpret the slope of the model using appropriate units of measure. For example, the amount spent is in dollars per month. Test the slope of the model. Report the 95% confidence interval estimate of the slope and explain what it represents. Find the Coefficient of Determination, R2, for your regression model and interpret the meaning. Estimate the average amount spent on streaming services for customers with a profile below. Keep in mind your regression in this part has YOUR selected x variable, so you will only use one of the values listed below to estimate the cost of monthly streaming service. 40 years of age, with a household income of $2200, and with a family size of 3. Part 2 Heres a thought: Could you find a better regression model based on your data? How about considering a multiple regression model which predicts the average amount spent on streaming services using ALL the data. Use Excels Analysis Toolpak to create your multiple regression model output. State the multiple regression model clearly by using the names of your variables in your equation. COST = _________ + _________(AGE) + _________(INCOME) + ____________(FAMILY SIZE) Be sure to include the regression output in your report. Test each of the slopes using the t-test. List these and write your conclusion for each. List each slope and give the value, test statistic, p-value, and conclusion of your test. Coefficient Test Statistic p-value Conclusion Age Income Family Size 3. Test the overall module using the F-Test statistic and related p-value. What is your conclusion from this test? 4. Report the Multiple Coefficient of Determination, R2, and interpret the meaning. 5. Use this model to estimate the average amount spent on streaming services for the profile: 50 years of age, with a household income of $3000, and with a family size of 4 6. Decide which model is best, the simple linear from Part 1 or the multiple regression from Part 2. Justify your choice. Submit a Word document of your written report. Use complete sentences and thoughts. Consider writing a paragraph for each part of this assignment. Report in your Word document all your findings using the calculations, tables, and graphics from Excel. You can copy and paste from Excel into Word. Create a report worthy of submission to management. Submit your Excel file demonstrating your work using the appropriate Excel commands or Excel Calculators for full credit.

    Attached Files (PDF/DOCX): MO7 Printable Business Application and Rubric.docx

    Note: Content extraction from these files is restricted, please review them manually.

  • Statistics Question

    Hello,

    I hope you are doing well.

    I am writing to request assistance with the SPSS assignment. (Plesse you should have the SPSS)

    All details are in the Word file and the ODE_building_high_school file in the link

    Please follow the rubric and ensure your answer is correct.

    Requirements: 2 days

  • Touchstone 5

    MAT 240 Module Five Assignment Guidelines and Rubric Scenario You have been hired by the Regional Real Estate Company to help them analyze real estate data. One of the companys Pacific region salespeople is working to design a new advertisement. The initial draft of the advertisement states that the average cost per square foot of home sales in the Pacific region is $280. The salesperson claims that the average cost per square foot in the Pacific region is less than $280. He wants you to make sure he can make that statement (that the average cost per square foot is less than $280) before asking for the advertisement text to be changed. In order to test his claim, you will generate a random sample size of 750 using data for the Pacific region and use this data to perform a hypothesis test. Prompt Generate a sample size of 750 houses using data for the Pacific region. Then, design a hypothesis test and interpret the results using significance level = .05. You will work with this sample in this assignment. Briefly describe how you generated your random sample. Use the House Listing Price by Region document to help support your work on this assignment. You may also use the Descriptive Statistics in Excel PDF and Creating Histograms in Excel PDF tutorials for support. Specifically, you must address the following rubric criteria: Introduction: Describe the purpose of this analysis and how you generated your random sample size of 750 houses. Hypothesis Test Setup: Define your population parameter, including hypothesis statements, and specify the appropriate test. o Define your population parameter. o Write the null and alternative hypotheses. o Specify the name of the test you will use. o Identify whether it is a left-tailed, right-tailed, or two-tailed test. Data Analysis Preparations: Describe sample summary statistics, provide a histogram and summary, check assumptions, and identify the test significance level. o Provide the descriptive statistics (sample size, mean, median, and standard deviation). o Provide a histogram of your sample. o Summarize your sample by writing a sentence describing the shape, center, and spread of your sample. o Check whether the assumptions to perform your identified test have been met. o Identify the test significance level. For example, = .05. Calculations: Calculate the p value, describe the p value and test statistic in regard to the normal curve graph, discuss how the p value relates to the significance level, and compare the p value to the significance level to reject or fail to reject the null hypothesis. o Calculate the sample mean and standard error. o Determine the appropriate test statistic, then calculate the test statistic. Note: This calculation is (mean target)/standard error. In this case, the mean is your regional mean (Pacific), and the target is 280. o Calculate the p value using one of the following tests. Choose your test from the following: =T.DIST.RT([test statistic], [degree of freedom]): right-tailed test =T.DIST([test statistic], [degree of freedom], 1): left-tailed test =T.DIST.2T([test statistic], [degree of freedom]): two-tailed test Note: The degree of freedom is calculated by subtracting 1 from your sample size. o Using the normal curve graph as a reference, describe where the p value and test statistic would be placed. Test Decision: Compare the relationship between the p value and the significance level, and decide to reject or fail to reject the null hypothesis. o Compare the relationship between the p value and significance level. o Decide to reject or fail to reject the null hypothesis. Conclusion: Discuss how your test relates to the hypothesis and discuss the statistical significance. o Explain in one paragraph how your test decision relates to your hypothesis and whether your conclusions are statistically significant. You can use the following tutorial that is specifically about this assignment: MAT-240 Module 5 Assignment Video What to Submit Module Five Assignment Template: Use this template to structure your report and submit the finished version as a Word document. Also submit your Excel file showing all steps and calculations used in the report.

    Attached Files (PDF/DOCX): MAT 240 Module Five Assignment Template.docx, week 5 mat240.docx, MAT 240 Creating Histograms in Excel.pdf, MAT 240 Descriptive Statistics in Excel(1).pdf, MAT 240 Descriptive Statistics in Excel(1).pdf, MAT 240 Creating Histograms in Excel.pdf, week 5 mat240.docx

    Note: Content extraction from these files is restricted, please review them manually.

  • discussion 6

    Results from surveys or opinion polls often report a range of valuesthe sample statistic plus or minus a margin of error (the resulting range is called a confidence interval). This tells us that the range is likely to contain the population parameter. How much wiggle room we provide is based on how much confidence we wish to have that the range contains the actual population mean. That confidence level is directly related to the middle “truth” area we will accept versus the dubious tail area we will rejectalso known as alpha (). The more confidence we wish to havethe more middle ground we will need to accept (more wiggle room)thus a smaller tail area. If we insist on a larger alpha (more dubious tail area), we narrow the middle ground we will accept and thus provide less wiggle roomso the more likely it is that we will miss the true average (and, thus, we have a lower confidence level). A 95% confidence level leaves 5% alpha. A 99% confidence level leaves 1% alpha.

    Now, without calculating a mean or margin of error or a confidence level, provide an example from your current (or your future) professional or personal life that describes a measurement that is normaland how much wiggle room on either side would be appropriate. When would you want a 95% confidence interval, and when would you want a 99% confidence level (a little more wiggle roomso a wider range)? This serves as your initial post to the discussion (if you choose topic 1) and is due by 11:59 p.m. EST on Saturday.

  • Statistic for Social Work

    Purpose

    Understanding and interpreting frequency distributions is a critical skill for social workers engaged in quantitative research and data analysis. Frequency distributions allow us to organize data in a meaningful way, identify patterns, and begin drawing inferences that can inform interventions, programs, and policy decisions.

    InstructionsInitial Post

    Using your own words, explain what a frequency distribution is and why it is a foundational tool in statistical analysis. Discuss the ways in which frequency distributions can be used to describe data, including references to shape (e.g., normal, skewed), central tendency, and variability.

    Choose a social work-related dataset or scenario (real or hypothetical) and describe how you would use a frequency distribution to interpret the data. For example, you might explore the frequency of adverse childhood experiences among youth in foster care, or the distribution of case closures by intervention type in a community agency.

    • What insights could a frequency distribution offer in your example?
    • What are the limitations of using frequency distributions alone to describe data?
    • How might these interpretations affect decisions made by social workers, administrators, or policymakers?

    Incorporate at least one scholarly source to support your response, and be sure to reflect on how this knowledge can enhance ethical and evidence-informed practice.

  • Week 4 Responses QR

    Respond to my colleague post separately:

    Colleague post#1: Listed below is my response to this weeks discussion question. I look forward to your comments and questions.

    Null hypothesis testing tells the researcher that something has occurred and it is significant. However, it can not tell us exactly what occurred. It can give the researcher the ability to state that what occurred did not occur by chance. But the reseracher can not state specifically what occurred. When testing to discredit the null hypothesis, flawed research can result if the researcher conducts the test incorrectly, thereby leading to Type I and Type II errors. Schmit (1996) argues that relying on significance testing has systematically retarded the growth of cumulative knowledge in psychology (p.115), while Cohen argues that testing for the null hypothesis encourages misinterpretation, misuse, and overconfidence, resulting in poor scientific judgment and distorted conclusions across studies (p.1308). These two articles got me thinking about what else would be affected if the findings were incorrect? This is where my past classes came in, and I began to examine threats to validity and reliability, specifically focusing on how flawed research contributes to compromises in the validity and reliability of the study findings. In research, there are two forms of validity; external validity and internal validity. External validity refers to how findings in one study could apply to a population in another study (Pearl & Bareinboim, 2024). This is known as generalizability (Pearl & Bareinboim, 2024). Thus, if the calculations in a study are done incorrectly, a researcher could make the argument that the findings do not apply to other areas because of x, y, z. Internal validity measures how well a study is conducted (its structure) and how accurately its results reflect the studied group (Cuncic, 2025, para.2). The same argument applies here, if the statistical tests are not done correctly, then the results will not be accurate, and will not reflect the studied group (Cuncic, 2025, para.2). Reliability refers to the consistency and reproducibility of measurements. It assesses the degree to which a measurement tool produces stable and dependable results when used repeatedly under the same conditions (McLead, 2024, para.1). When reporting the statistical testing section of research this is where the researcher says, x, y, z test was used to conduct a, b, c. This test is deemed reliable because it has been used in x number of tests with y consistent findings. To the average person, this sounds valid. However, the average person does not consider how the test was conducted, what the population was, or whether the study was conducted correctly. I teach this concept in my business classes using late-night infomercials as the example, specifically talking about how statistics and statements are skewed to sell a product. I tell my students, before you go out and buy the latest thing you saw on late-night TV, ask yourself this question: Can I find the research results they are talking about? If not, do not buy the product because it is most likely a gimmick, and they are using statistics to make their product sound legit.

    References

    Abos, P. (2024). Validity and Reliability: The extent to which your research findings are accurate

    and consistent. Retrieved February 4, 2026, from

    https://www.researchgate.net/publication/384402476_Validity_and_Reliability_The_

    extent_to_which_your_research_findings_are_accurate_and_consistent

    Cohen, J. (1990). Things I Have Learned (So Far). American Psychologist 45(12), 1304-1312.

    Cuncic, A. (2025). Internal Validity vs. External Validity in Research. Retrieved February 4, 2026,

    from https://www.verywellmind.com/internal-and-external-validity-4584479

    MLead, s. (2024). Reliability vs. Validity in Research. Retrived Fevruary 4, 2026, from

    Reliability vs Validity in Research

    Pearl, J., & Bareinboim, E. (2022). External Validity: From Do-Calculus to Transportability Across

    Populations. ACM Books.

    Schmidt, F., L. (1996). Statistical Significance Testing and Cumulative Knowledge in Psychology:

    Implications for Training of Researchers. Psychology Methods, 1(2), 115-129.

    Colleague post #2: The Major Flaws of Null Hypothesis Significance Testing

    Null hypothesis significance testing (NHST) has been the dominant framework in psychological research for decades, yet many scholars argue that it contains fundamental flaws that limit the fields ability to build reliable and cumulative scientific knowledge. Two of the most influential critiques come from Cohen (1990) and Schmidt (1996), both of whom highlight how NHST is frequently misunderstood, misapplied, and overvalued in psychological science (Cohen,1990; Schmidt,1996 )

    Cohen (1990) argues that researchers routinely misinterpret the meaning of the pvalue, treating it as a direct indicator of truth rather than a conditional probability based on hypothetical repeated sampling. He emphasizes that statistical significance does not imply practical significance, noting that trivial effects can become significant with large enough samples (Cohen,1990 ). Cohen also points out that NHST encourages researchers to ignore effect sizes and statistical power, which are essential for understanding the magnitude and reliability of findings. His critique suggests that the field often mistakes detectability for importance, leading to a literature filled with statistically significant but scientifically uninformative results (Cohen,1990; Schmidt,1996 )

    Schmidt (1996) extends this critique, arguing that NHST actively prevents psychology from developing cumulative knowledge. Because NHST focuses on binary decisions, significant or not, its approach obscures the true size and consistency of effects across studies (Schmidt,1996 ). Schmidt contends that this dichotomous thinking contributes to publication bias, unstable findings, and a lack of theoretical progress. He also notes that NHST is overly sensitive to sample size, producing significant results for trivial effects in large samples and nonsignificant results for meaningful effects in small samples (Schmidt,1996 ). According to Schmidt, the belief that NHST provides objective scientific rigor is illusory, and the field must shift toward effect sizes, confidence intervals, and meta-analytic thinking to advance (Schmidt, 1996).

    Together, Cohen and Schmidts critiques reveal that NHST is not merely a flawed statistical tool but a barrier to scientific progress when used uncritically. Their work has helped push psychology toward more informative approaches, such as estimation statistics, effect size reporting, and metaanalysis, that better support cumulative knowledge and theoretical development. As the field continues to confront issues such as the replication crisis, these critiques remain highly relevant and underscore the need for statistical practices that prioritize meaning over mere significance (Cohen, 1990; Schmidt, 1996). During the replication crisis, another author, Bargh et al. (1996), did not directly critique NHST, but their study became one of the most influential examples of the flaws that Cohen (1990) and Schmidt (1996) warned about. The priming effects reported by Bargh were statistically significant but later failed to replicate, illustrating how NHST can produce unstable findings, encourage overreliance on pvalues, and hinder cumulative scientific progress. Replication failures (e.g., Doyen et al., 2012) provide empirical support for Cohens and Schmidts critiques (Cohen,1990; Bargh et al.,1996; Schmidt,1996;D oyen et al.,2012 )

    What do these flaws mean for the field of Psychology? The flaws mean as follows. Psychology risks building theories on statistical significance is driven by sample size rather than meaningful effects, the field may chase’significant’ but trivial findings. NHST contributes to the replication crisis. Low power, overreliance on p-values, and publication bias all contribute to poor replicability, one of the biggest issues in modern psychology. The field needs to shift toward estimation and cumulative science (Cohen, 1990; Schmidt, 1996). Both Cohen and Schmidt advocate for effect sizes, confidence intervals, Meta-analysis, power analysis, and transparent reporting. NHST should not be the primary decision-making tool. Neither Cohen nor Schmidt argued for eliminating NHST entirely, but both insisted it should be supplemented or replaced by more informative statistical approaches. Finally, Cohen and Schmidt both argue that NHST is fundamentally limited and misleading. Their critiques helped spark the modern movement towards effect sizes, confidence intervals, meta-analysis, and estimation-based statistics, approaches that provide richer, more meaningful information than a simple p-value ( Cohen,1990; Schmidt,1996 )

    Summary Table

    Issue Cohen (1990 ) Schmidt (1996)

    Misinterpretation of p-value yes yes

    No statistical significance, practical

    Significance yes yes

    Low power in psychology yes

    NHST blocks cumulative knowledge yes

    Dichotomous thinking yes

    Sample size distortions yes yes

    Need for effect sizes & power yes yes

    References

    Bargh, J. A., Chen, M., & Burrows, L. (1996). Automaticity of social behavior: Direct effects of trait construct and stereotype activation on action. Journal of Personality and Social Psychology, 71(2), 230244. American Psychological Association

    Doyen, S., Klein, O., Pichon, C.L., & Cleeremans, A. (2012). Behavioral priming: Its all in the mind, but whose mind? PLOS ONE, 7(1), e29081.Public Library of Science DOI: 10.1371/journal.pone.0029081

    Cohen, J. (1990). Things I have learned (so far). American Psychologist, 45(12), 13041312.

    American Psychological Association

    DOI: https://doi.org/10.1037/0003-066X.45.12.1304

    Schmidt, F. L. (1996). Statistical significance testing and cumulative knowledge in psychology: Implications for training of researchers. Psychological Methods, 1(2), 115129.

    American Psychological Association

    DOI: https://doi.org/10.1037/1082-989X.1.2.115.

    Colleague post #3: Null hypothesis significance testing has several major flaws that have long been recognized as problematic for psychological science. One key issue, emphasized by Cohen (1990), is that statistical significance is often misunderstood as evidence that an effect is important or meaningful, when it only indicates that an effect is unlikely to be zero given the sample size. Because the null hypothesis almost always states that an effect is exactly zero, a condition that is rarely true in real-world psychological phenomena, researchers are often rejecting a hypothesis that was never plausible to begin with. As a result, null hypothesis significance testing encourages a misleading focus on p values rather than on the size of effects, their practical importance, or the theoretical meaning of the findings. This has led to a culture in which results are reduced to binary decisions (significant vs. not significant), oversimplifying complex psychological processes.

    Another serious problem, highlighted by Schmidt (1996), is that reliance on significance testing actively slows the development of cumulative knowledge in psychology. Because null hypothesis significance testing focuses heavily on controlling Type I error (false positives), it often ignores Type II error (false negatives) and statistical power. This leads to many real effects going undetected, especially in studies with small sample sizes. Schmidt shows that entire research literatures can appear contradictory simply because some studies reach statistical significance and others do not, even when all are estimating the same underlying effect. This vote-counting approach creates the false impression that effects are inconsistent or unreliable, when in fact the inconsistency is largely due to sampling error and low power rather than true differences in psychological phenomena.

    Together, Cohen (1990) and Schmidt (1996) argue that the field of psychology has been misled by overreliance on null hypothesis significance testing and that this has serious consequences for theory building, replication, and practical application. Both authors emphasize that researchers should shift their focus toward effect sizes, confidence intervals, and meta-analysis, which provide more informative and honest summaries of research findings. Without this shift, psychology risks continuing to produce fragmented and confusing research literatures that obscure rather than clarify real effects. Moving beyond simple significance testing is therefore essential for advancing cumulative knowledge and improving the scientific credibility of the field.

    References

    Cohen, J. (1990). Things I have learned (so far). American Psychologist, 45(12), 13041312.

    Schmidt, F. L. (1996). Statistical significance testing and cumulative knowledge in psychology: Implications for training of researchers. Psychological Methods, 1(2), 115129.

  • Upload the practice

    Upload the practice with your work or R file if you decide to use R.

    Requirements: 4 hours

  • Upload the practice

    Upload the practice with your work or R file if you decide to use R.

    Requirements: 3 hours

  • STATS assignment

    Watch the TED talk by Arthur Benjamin. Write an open-ended reflection on the video. Questions you may choose to consider include:

    • If you have taken calculus, have you had chances to use in in your everyday life or career? Can you think of types of problems where calculus is useful?
    • In a similar vein, can you think of problems or careers where statistics is useful?
    • Arthur mentioned that statistics is used in building games, analyzing trends, and predicting the future. How does probability factor into these types of problems?
    • Do you agree with Arthur Benjamin’s argument? Why or why not?

    No specific source is required just the TED talk video