Legislative Critique of the Maritime Transportation Security Act 2002 and the S.A.F.E. Port Act of 2006

Maritime Transportation Security Act 2002
Maritime Transportation Security Act 2002

Want help to write your Essay or Assignments? Click here

Legislative Critique of the Maritime Transportation Security Act 2002 and the S.A.F.E. Port Act of 2006.

There are two significant pieces of legislation that impact maritime security in the United States. These include the Maritime Transportation Security Act 2002 and the S.A.F.E. Port Act of 2006. 

For this assignment, you are to conduct a graduate level Legislative Critique of these two pieces of legislation. To do this critique perform the following steps:

. Obtain a copy of the piece of legislation to be examined. (MTSA and S.A.F.E. Port Act of 2006)

2. Collect and analyze news media reports on the legislation.

3. Collect and analyze any other critiques of the legislation. (i.e. Think tanks, academics, etc)

4. Collect and analyze available witness testimony with regard to the legislation.

To compose a legislative critique students should compose an analysis based upon the following points: 

• Provide a brief legislative history of the Bills. (No more that 500 words – USE WORD ECONOMY)!

• Determine if the Bills focus was introduced in the past under a different name and/or format. If so, evaluate why the legislation was discarded. 

• Summarize the main points of each bill in your own words (but supported by citation as appropriate)

• Evaluate and categorize the strengths and weaknesses of the individual legislation, then compared against each other. 

• Conclude with a final assessment of both pieces of legislation including SPECIFIC recommendations that could be considered to enhance the security framework that has been established.

These recommendations need to be as specific as possible JUSTIFY why you believe this should be considered.
MTSA- https://www.congress.gov/107/plaws/publ295/PLAW-107publ295.pdf
S.A.F.E. Port Act of 2006- https://www.congress.gov/109/plaws/publ347/PLAW-109publ347.pdf

Want help to write your Essay or Assignments? Click here

Sampling Distributions and Confidence Intervals

Sampling Distributions and Confidence Intervals
Sampling Distributions and Confidence Intervals

We can write this or a similar paper for you! Simply fill the order form!

Sampling Distributions and Confidence Intervals

ORDER INSTRUCTIONS:

1. What is the central limit theorem? What role does it play in the establishment of the sampling distribution of the sample mean? 
2. Contrast normal distributions with sampling distributions.

In statistical inference, one wishes to estimate population parameters using observed sample data.

A confidence interval gives an estimated range of values which is likely to include an unknown population parameter, the estimated range being calculated from a given set of sample data. (Definition taken from Valerie J. Easton and John H. McColl’s Statistics Glossary v1.1)

The common notation for the parameter in question is . Often, this parameter is the population mean , which is estimated through the sample mean .

Below is a partial answer to the above homework questions by one of our writers. If you are interested in a custom non plagiarized top quality answer, click order now to place your order.

Sampling Distributions and Confidence Intervals Essay

The central limit theorem

The central limit theorem refers to a theory in probability statistics where the approach is that the distributions of a normalized….

The role of the Central Limit Theorem in the establishment of the sampling distributions of the sample mean

A major role of the central Limit theorem in terms of establishing sampling distributions is the use of inference from the available….

Contrasting normal distributions with sampling distributions

Normal distributions refer to a normal continuous probability distribution that when….

We can write this or a similar paper for you! Simply fill the order form!

R studio Assignment Paper

R studio
R studio

R studio

Want help to write your Essay or Assignments? Click here

R studio

Order Instructions:

The data for this week’s assignment is the Breast Cancer data set found in the mlbench library for computing in the R studio

It consists of 699 observations on 11 variables. The names of the variables are (check the help information for the data for some other information).

Id, the code number (not used in the analysis)

cl.thikness, the clump thickness

cell.size, cell size

cell.shape, cell shape

Marg.adhesion, marginal adhesion

Epith.c.size, single epithelial cell size

Want help to write your Essay or Assignments? Click here

Bare.nuclei, bare nuclei

Bl.cromatin, Bland cromatin

Normal.nucleoli, normal nucleoli

Mitoses

Class (this is the target variable)

The goal is to predict values of Class.

Variables 2 through 10 are coded as factors: for some of the work you may want to convert them to numerical values and standardize them.

Want help to write your Essay or Assignments? Click here

In your analysis

Examine the data for missing values, and explain how you decided to deal with them

Examine the distributions of variables 2 through 10 (boxplots or histograms, or both) for shape and outliers. These variables are recorded so that all are scaled to a range of 0 to 10, but their distributions may not be (probably are not) the same, hence the need for investigation. If you want to create histograms, note that the hist() function in R requires a vector of values, so you will need to do something like 
hist(as.numeric(BreastCancer$cell.size))
to get the histogram to work

Prepare the data for use in a neural network

Fit a single hidden layer neural network to the data and assess its performance. (Do the usual training/testing process). Try to tune the model to select an ‘optimal’ value of hidden nodes

Use network averaging to model the data, and assess its performance

We can write this or a similar paper for you! Simply fill the order form!

FTE and Variance Assignment

FTE and Variance
FTE and Variance

FTE and Variance

Want help to write your Essay or Assignments? Click here 

Section One: Calculation of Full-Time Equivalents

Personnel Budget Case Study

Question One

There are a number of items that are required in order to ensure that a budget is effectively prepared. According to lectures of this course, several items that are required include: patient acuity, patient days, length of a patient stay, number of the vacancies that are anticipated in each level, staffs’ educational needs in the in the year that follows, costs associated with staff benefits, as well as non-productive staff time that is anticipated. The background information/data about 1 West unit is as follows:

Patient Data: Average Daily Census (30); Unit Capacity (32); Average HPPD (8.8); and Total Care Hours (96,360)

Staff Data: Productive hours/employee/year (1,780); Nonproductive hours/employee/year (300); and Total Hours/employee/year (2,080)

Skill Mix: RNs (80%); LVNs (10%); and Nurse Aides (10%)

Want help to write your Essay or Assignments? Click here

Question Two

Calculating the number of the productive FTEs to be needed:

The initial step would be to consider the subsequent year’s workload, figuring in mind that the anticipated number of patient days will be 10,950 [obtained by multiplying average daily census (30) by 365 days].

Step 1:

Calculating Workload = HPPD x Number of patient days = 8.8 x 10,950 = 96,360

Step 2:

Calculating Productive time = Subtracting total nonproductive hours from total FTE hours; 2,080 – 300 = 1,780 hours

Step 3:

Calculating FTEs = Workload divided by productive time: 96,360/1,780 = 54.3 productive FTEs needed to staff the unit.

Want help to write your Essay or Assignments? Click here

Question Three

Determining the number of RNs, LVNs, as well as nurse aides that will be required to ensure that the unit is well staffed based on the assumption that there will be a 12-hours working shifts for the staff. In other words, this requires the calculation of the number of persons-shifts to be need during periods of 24-hour shifts.

Since the staffs are shared 50% for day shift and 50% for night shift in the following mix RNs (80%); LVNs (10%); and Nurse Aides (10%), the numbers are as follows:

In a 12-hour shift:

RNs = 80/100 * 54.3 = 43.44

LVNs = 10/100 * 54.3 = 5.43

Nurse aides = 10/100 * 54.3 = 5.43 

Want help to write your Essay or Assignments? Click here

Question Four

Assigning the staff by shift and by type considering that:

Day Shift                    50%

Night Shift                  50%

Then,

RNs = 80/100 * 54.3 = 43.44 each shift

LVNs = 10/100 * 54.3 = 5.43 each shift

Nurse aides = 10/100 * 54.3 = 5.43 each shift

Want help to write your Essay or Assignments? Click here

Question Five

Coverting staff positions to full-time Equivalents or FTE positions involves converting to 24/7 as follows:

Total FTEs = 54.3 x 1.4 = 76.02 FTEs needed to staff the unit 24/7, which is equivalent to 76.02 FTE positions

This will help convert staff positions to FTE positions needed for a 24/7 shift as follows:

RNs = 80/100 * 76.02 = 60.8

LVNs = 10/100 * 76.02 = 7.6

Nurse aides = 10/100 * 76.02 = 7.6

Want help to write your Essay or Assignments? Click here

Section Two: Variance Analysis

Variance Analysis Case Study

  1. Comparing the original budget to the flexible budget reveals that, the original one had no significant unfavorable variances compared to the flexible one, which is characterized by unfavorable variances not only in terms of volume but also in terms of price and quantity.  For instance, a volume variance is experienced when the actual volume is higher or lower than the budgeted volume and may be expressed in terms of FTEs or patient days.  To determine volume variance, the calculation is as follows:

 (Budgeted Volume – Actual Unit Volume)(Budgeted Rate) = Volume Variance

In order to get the budgeted rate, the following formula is used:

Budgeted Rate = Budget Allocation/Budgeted Volumes

Thus,

Budgeted Rate = $40 *5 * 340/340 = $200

Volume Variance = (340 – 400) (200) = – 12,000

Want help to write your Essay or Assignments? Click here

A comparison between the flexible budget and actual budget shows that, there are unfavorable variations in terms of price and quantity. This is because employees’ average hourly rate has increased from $40.00 to $45.00, whereas hours per care per patient have increased from 5.0 to 5.6. These two changes are indicative of an unfavorable variation. Price variance or unit cost variance determination can be calculated using the equation shown below:

 (Budgeted Unit Price – Actual Unit Cost)(Actual Volume) = Unit Cost Variance

Price variance = [(40 * 5) – (45*5.6) (400)] = – $20,800

In addition, quantity variance determination can be calculated using the equation shown below:

 (Budgeted Use – Actual Use)(Budgeted Unit Cost) = Quantity Variance

 Quantity variance = [(340 – (400) (40 * 5)] = -12,000   

Want help to write your Essay or Assignments? Click here

  • There are several factors that led to the differences in variances including both external and internal factors. In particular, the internal factors include changes in staff efficiency which has led to an increase in the hours per care per patient from 5.0 to 5.6. Changes in technology as well as nature of surgeries may have also be other internal factors that have led to these variations.
  • This is because all of these factors can collaboratively combine to delay the rate at which patients are operated eventually increasing the overall average time required to provide care to each patient in hours per visit. On the other hand, external factors that may have caused the variations include type of staff available, census changes as well as price changes, all of which can be attributed to an increase in the hours per care per patient from 5.0 to 5.6, number of visits from 340 to 400 as well as employees’ average hourly rate from $40.00 to $45.00. The unfavorable variations occurred because the above discussed factors combined to negatively impact the budgeted figures.

References

Zimmermann, P. G. (2002). Nursing management secrets, Issue 974; Volume 13 of Secrets series. Elsevier Health Sciences. p. 55. 

Want help to write your Essay or Assignments? Click here

Statistics in Prediction of Outcome

Statistics in Prediction of Outcome
Statistics in Prediction of Outcome

Want help to write your Essay or Assignments? Click here

Statistics in Prediction of Outcome

Introduction

Statistic can be used to predict outcome. This is because statistic encompasses the use of various techniques that enable a person in prediction of outcome. To emphasize this fact, we carried out a project. Our project was about analyzing data to extend where we answer the research question. The topic was chosen because there is need to understand the relationship between statistic and outcome. We wanted to come up with proper understanding of the reliability of statistic in prediction of outcome. The research question was ‘can statistic predict income?’

Saint Leo University has excellence as one of the core values. The excellence that the University dreams of is about creating character that enables each one of the students to end up responsible people in the society. This is brought about by team work, dedication and completion of assigned tasks. Therefore, completing this project reflected what the value of excellence at Saint Leo University advocates.

Data

The data that we used is secondary. The data used in this report was obtained from the website of Halliburton Limited, a multinational company operating in the oil fields service providers industry.

 This data covers a period of 3 years, from the year 2012 to 2014 and was obtained from http://ir.halliburton.com/phoenix.zhtml?node-id=huzgm7u7&c=67605&p=irol-reportsOther

Variables

In this project, we used dependent and independent variables. Dependent variables are the ones that a researcher is always interested in. On the other hand, the independent variables are the ones that affect the dependent variables.

Want help to write your Essay or Assignments? Click here

Dependent and independent variables

In our project, the dependent variables consist of operating income for three years. On the other hand, the independent variables consist of sales values for three years. According to Halliburton (2014), the sales and operating income values were as below.

Year Sales (x) Profit(y)
2012 28,503.00     4,159.00
2013 29,402.00     3,138.00
2014 32,870.00     5,097.00

Descriptive Statistics Table

Statistic Sales (x) Profit(y)
Mean   30,258.33   4,131.33
Median   29,402.00   4,159.00
Min   28,503.00   3,138.00
Max   32,870.00   5,097.00
Standard Deviation1882.84799.998
Skewness1.36           0.10

Statistic Test

The testing that we carried out is Pearson coefficient of correlation. This test is quite useful in understanding the relationship of variables. It is very useful in understanding how strong the relationship between variables is.

Year Sales (x)Operating Income (y) (xy) (x^2) (y^2)
2014 32,870.00    5,097.00  167,538,390.00  1,080,436,900.00  25,979,409.00
2013 29,402.00    3,138.00    92,263,476.00     864,477,604.00    9,847,044.00
2012 28,503.00    4,159.00  118,543,977.00     812,421,009.00  17,297,281.00
Total 90,775.00 12,394.00  378,345,843.00  2,757,335,513.00  53,123,734.00

The formulae of Pearson Coefficient of correlation is as shown below

N                            3.00
 N∑xy   1,135,037,529.00
∑x∑y   1,125,065,350.00
  
 N∑x^2)   8,272,006,539.00
 ∑x^2)   2,757,335,513.00
 N∑y^2)       159,371,202.00
 ∑y^2)         53,123,734.00

Coefficient of correlation (r) = 1,135,037,529.00-1,125,065,350.00/ Square root of (8,272,006,539.00-2,757,335,513.00)( 159,371,202.00- 53,123,734.00)

= 9,972,179 / 292,959,920,000,000,000.00

Coefficient of correlation (r) = 0.00000000003

After conducting Pearson coefficient of correlation, the result was a positive correlation of the variables. This means that when sales increase, the operating income increases too. This shows that we can predict the outcome of the year 2015.

Conclusion

From the project, it is evident that statistic can be used to predict outcome. Therefore, the research question has been addressed well. From the analysis, it has become evident that the use of statistic is faced with certain challenges. Firstly, reliability of some data sources is questionable. Additionally, the fact that data is not readily available is another challenge. There has also been an important observation that the choice of variables should be done properly to avoid confusion of the tests carried out. From the project, it has come out clearly that when there is positive correlation of variables, the movement of change in any of them is positive.

References

Halliburton (2014), Hallinurton, Retrieved from http://ir.halliburton.com/phoenix.zhtml?node-id=huzgm7u7&c=67605&p=irol-reportsOther Last Accessed 19th April, 2016

Want help to write your Essay or Assignments? Click here

Generating a Hypothesis

Generating a Hypothesis
Generating a Hypothesis

Inferential Statistics and Findings: Generating a Hypothesis

The aim of the paper is to generate a hypothesis with the use of research questions and two variables- dependent variable: sales/revenue and independent variable: interest rates. In addition, the paper uses a statistical technique for testing the hypothesis at 95% confidence interval while interpreting the results. XYZ Corporation is a US-based company that majors in home improvement and construction.

Moreover, the company operates in various big-box format stores across the United States. Nevertheless, XYZ Corporation wants to determine the effects of interest rates and sales on its future profitability, therefore, the company approached a research team that suggested the below research question.

Research question

Is there a relationship between federal (Feds) directed interest rate (ID) and sales /revenue (DV) on future profit of XYZ Corporation?

Hypothesis
H0 = XYZ Corporation’s future profits are not determined by interest rates and sales/revenue
H1 = XYZ Corporation’s future profits are determined by interest rates and sales/revenue

Hypothesis tests with a 95% confidence level

The veracity of the null hypothesis depends on data  as well as the objective of significance tests to estimate the likelihood, hence the P-value, which underscores that the occurrence is an issue of chance. In this case, if the Pearson Correlation is less than 5% it means that  the null hypothesis is rejected.

However, if the P-value is greater than 5% then it means the statistical test fails to reject the null hypothesis (Gupta, 2012).  Data was indiscriminately selected from 382 observations with respect to interest rates and sales of XYZ Corporation.

Interpretation of the results and findings

            Statistical analysis demonstrates that this is a normal distribution, with a mean of 10.97644 and standard deviation of 10.14628 for sales (Table 1). In this case, sales were observed between $0.00 and $89.00.  Changes in the federal interest rates could help determine sales of XYZ Corporation.

In addition, the mean and standard deviation for interest rates were observed at 9.695538 and 7.836266 respectively (Table 2).  High-interest rates are highly likely to inhibit any profit making from lack of sales.  While low-interest rates by the federal government may lead to increased sales. 

Again, it demonstrated a 95% confidence interval (sales:1.020716, interest rates: 0.789369) this implies that XYZ Corporation can be about 95 percent confident that sales and interest rates will lead to an increased future a profit. Because the confidence interval of sales corresponds to the value of alpha, which is 1 indicates that there is no significant correlation between sales and future profit of XYZ Corporation.

However, the results are empirically reliable. When it comes to interest rates the confidence interval is less than 1, statistically the test fails to refute the null hypothesis, an aspect that demonstrates statistical significance of the test (Monette, Sullivan & DeJong, 2011).

Table 1. Descriptive statistics showing XYZ Corporation Sales

Sales
Mean10.97644
Standard Error0.519129
Median8
Mode5
Standard Deviation10.14628
Sample Variance102.947
Kurtosis18.50581
Skewness3.065591
Range89
Minimum0
Maximum89
Sum4193
Count382
Confidence Level (95.0%)1.020716

Table 2. Descriptive statistics showing XYZ Corporation Interest Rates

Mean9.695538
Standard Error0.401464
Median8
Mode8
Standard Deviation7.836266
Sample Variance61.40706
Kurtosis27.77302
Skewness3.393958
Range89
Minimum0
Maximum89
Sum3694
Count381
Largest(1)89
Smallest(1)0
Confidence Level (95.0%)0.789369

References

Gupta, S. K. (2012). The relevance of confidence interval and P-value in inferential. Indian Journal of Pharmacology. Retrieved on  June 9, 2016, from statistics http://www.ncbi.nlm.nih.gov/pmc/articles/PMC3271529/

Monette, D. R., Sullivan, T. J., & DeJong, C. R. (2011). Applied Social Research: A Tool for the Human Services (8th ed.). Belmont, CA: Brooks/Cole.

Want help to write your Essay or Assignments? Click here

Descriptive statistics and inferential statistics

Descriptive statistics and inferential statistics

Descriptive statistics and inferential statistics

  1. Introduction

This report is an analysis of the data generated using SPSS and presented using charts and tables. The report firstly presents the results of selected descriptive statistical analyses. Subsequently, the report summarises the numerical results with descriptive statistics analysis tables or graphs, including the interpretation of these tables and graphs. The fourth section or the report is a presentation of the data regarding numerical results of the inferential statistics. This is followed by a discussion of the same, before a summative conclusion is presented in the last section.

  • Selected descriptive statistics

Descriptive statistics refers to the kinds of data that analysts and researchers use in presenting the characteristics of the sample used in a study. According to Kothari (2004), they are used in checking whether the variables that the researcher has chosen to use violate any assumptions that the researcher might have made, which might be consequential to the findings. Another important function of descriptive statistics used in this section is that they help to answer the core research questions.

In the present study, the descriptive statistics selected are for public use micro data area code (PUMA), house weight (WHTP), state code, (ST), numbering of persons (NP), rooms (RMS), bedrooms, (BDS), and household income (HINCP). The data retrieved was as presented in table 1 below

Table 1: PUMA, ST, BDS, RMS, mean, median, and standard deviation

 RMSBDSSTPUMA
NValid4911491149114911
Missing0000
Mean4.872.6115.00248.05
Median5.003.0015.00302.00
Std. Deviation1.9331.197.00081.573
Minimum1015100
Maximum9515307

Table 2: RMS, BDS, ST, and PUMA, frequency table

 PUMAFrequencyPercentValid PercentCumulative Percent
Valid10095119.419.419.4
 20078215.915.935.3
 3014078.38.343.6
 3024128.48.452.0
 3034258.78.760.6
 30453610.910.971.5
 3053657.47.479.0
 3064569.39.388.3
 30757711.711.7100.0
 Total4911100.0100.0 
NPFrequencyPercentValid PercentCumulative Percent
Valid04529.29.29.2
 197019.819.829.0
 2149130.430.459.3
 371114.514.573.8
 461912.612.686.4
 53186.56.592.9
 61613.33.396.2
 7941.91.998.1
 830.6.698.7
 920.4.499.1
 1018.4.499.5
 117.1.199.6
 127.1.199.7
 135.1.199.8
 151.0.099.9
 162.0.099.9
 171.0.099.9
 192.0.0100.0
 202.0.0100.0
 Total4911100.0100.0 
RMSFrequencyPercentValid PercentCumulative Percent
Valid11853.83.83.8
 23457.07.010.8
 367713.813.824.6
 489618.218.242.8
 5111022.622.665.4
 676815.615.681.1
 74388.98.990.0
 82344.84.894.7
 92585.35.3100.0
 Total4911100.0100.0 
BDSFrequencyPercentValid PercentCumulative Percent
Valid02114.34.34.3
 168313.913.918.2
 2120824.624.642.8
 3181036.936.979.7
 468814.014.093.7
 53116.36.3100.0
 Total4911100.0100.0 

From the data in table 1 above, a number of observations are blatant and clear. The first is that the means of RMS, BDS, ST and PUMA are 4.87, 2.61, 15, and 248.05 respectively. For rooms, the number of rooms, the median score was 5, where the scores varied from 1 to 9. This means that the majority of respondents have about 5 rooms.

When it comes to the number of bedrooms, the median score was 3, whereas the mean was 2.61, this shows that the majority of respondents have 3 rooms. The state code was 15 for all respondents whereas the mean for public use of micro data area code was 248.05. The mean was 302, whereas the minimum and maximum scores were 100 and 307 respectively.

From table 2, a number of assertions can also be made, and the first is about PUMA. From the table, the evidence shows that for public use of micro data area code, 19.4% of the respondents scored category 100, which made it the highest selected category, whereas 15.9% of the respondents checked 200, making it the second most selected category. Comparatively, 301 was the least selected category at 8.7%.

Additionally, for number of bedrooms, a majority of the respondents said that they had three bedrooms in their houses, and this represented 3.9% of all responses, closely followed by those with two bedrooms at 24.6%. At the same time, the number of people living in houses with no bedrooms or five bedrooms was the least with a score of 4.3% and 6.3% respectively.

This data is in line with the data about rooms, which shows that 22% of respondents stay in a five-roomed apartment, followed by 18% and 15%, who stay in four and five roomed houses respectively. Because of the number of rooms and bedrooms in their houses, it is plausible to conclude that a majority of the respondents stay with other people or expect other people to visit often, which are why they have extra rooms in the house, as well as extra bedrooms in the house.

Additionally, from the data, it is obvious that a majority of the people are in the middle between the rich and the poor, as those who stay in studio apartments are as marginal as those who stay in luxury apartments that can contain at least five bedrooms. . 

  • Selected inferential statistical analyses

Inferential statistics refer to the data analysis methods where the researcher or analyst uses a given set of data to determine whether there is a link between given variables being studied. By using inferential statistics, the researcher can tell whether the relationship that seems to exist between variables is a fact, or whether it is not a fact. According to Kothari (2004), a number of measures and techniques can be used to accomplish inferential statistics. The two types of inferential statistics used in this report are correlation and regression analyses.  

Correlation was conducted using the Pearson correlation analysis. Pearson correlation analysis is employed to measure the linear relationship between two or more variables. The value of Pearson correlation ranges between -1 and +1, with -1 indicating negative correlation, 0 indicating no correlation and +1 indicating positive correlation between the variables.  Besides, the closer the value is to +1, the stronger the relationship between the variables (Saunders, Lewis & Thornhill, 2007). For this study, the data is as shown below.

According to table 4-20, Sig. (2-tailed) =0.000, and all the four variables have a significant correlation at the 0.01 significant level. Pearson correlation between PUMA and NP is .110, whereas the relation between PUMA and BDS and RMS is .042 and .067 respectively. This shows that there is a weak but positive relationship between PUMA and all the independent variables, although the weakest relationship is that between PUMA and BDS.

Table 3: Correlations

 PUMANPBDSRMS
PUMAPearson Correlation1.110(**).042(**).067(**)
Sig. (2-tailed) .000.003.000
N4911491149114911
NPPearson Correlation.110(**)1.447(**).396(**)
Sig. (2-tailed).000 .000.000
N4911491149114911
BDSPearson Correlation.042(**).447(**)1.878(**)
Sig. (2-tailed).003.000 .000
N4911491149114911
RMSPearson Correlation.067(**).396(**).878(**)1
Sig. (2-tailed).000.000.000 
N4911491149114911

**  Correlation is significant at the 0.01 level (2-tailed).

Regression analysis helps estimate and investigate the association between variables. R Square is used to show the degree of relationship between the dependent and independent variables. R Square value ranges between 0 and 1, and the closer the value is to 1, the stronger the relationship between the variables further indicating the greater degree to which variation in independent variable explains the variation in dependent variable (Seber and Lee, 2012).

Based on the model summary table 4-21, R stand for the correlation coefficient and it depicts the association between dependent variable and independent variables. It is evident that a positive relationship exists between the dependent variable and independent variables as shown by R value (0.126).

However, the relationship is a very weak one. Besides, it can be seen that the variation in the three independent variables (RMS, BDS and NP) explain 1.6% variation of PUMA as represented by the value of R Square. Therefore, it means that other factors that are not studied on in this study contribute 98.4% of the PUMA programs. This means that the other factors are very important and thus need to be put into account in any effort to enhance PUMA. Additionally, this research therefore identifies the three independent variable studied on in this research as the non-critical determinants of PUMA boundaries.

Table 4: regression analysis results

                                 Model Summary

ModelRR SquareAdjusted R SquareStd. Error of the Estimate
1.126(a).016.01580.945

a  Predictors: (Constant), RMS, NP, BDS

Further, this research established through the analysis f variance that the significant value is 0.00, which is less than 0.01, therefore the model is statistically significant in foretelling how NP, RMS, and BDS can influence PUMA groupings. The F critical value at the 0.01 level of significant was 26.501. Given that F calculated  is greater than the F critical value of 26.501, then it means that the overall model was significant (Seber and Lee, 2012).

                                                         ANOVA(b)

Model Sum of SquaresdfMean SquareFSig.
1Regression520911.1203173637.04026.501.000(a)
 Residual32151092.84549076552.087  
 Total32672003.9654910   

a  Predictors: (Constant), RMS, NP, BDS

b  Dependent Variable: PUMA

At the same time, the beta coefficients also gives significant inferential information. According to the regression coefficients presented in table 4-23, this research found that when all independent variables (the number of persons (NP), number of rooms (RMS), and the number of bedrooms (BDS)) are kept constant at zero, the level of public use micro data area code (PUMA)  will be at 231.13. A 1% change in number of persons will lead to an 11.4% increase in PUMA, whereas a one percent change in BDS will lead to a 12.1% changes in PUMA.

Comparatively, a one percent change in  RMS will lead to a 12.8 percent change in PUMA. This leads to the conclusion that of the three variavles, RMS leads to the largest impact in PUMA when the three independent variables are pitted together. Further the statistical significance of each independent variable was tested at the 0.01 level of significance of the p-values.

                                                      Coefficients(a)

Model Unstandardized CoefficientsStandardized CoefficientstSig.
BStd. ErrorBeta
1
(Constant)231.1303.161 73.128.000
NP4.700.654.1147.181.000
BDS-8.2222.068-.121-3.977.000
RMS5.3841.248.1284.315.000

a  Dependent Variable: PUMA

In general form, it can be said that the equation used to determine the link between  Public use microdata area code, numbering of persons, rooms and bedrooms is of the form:

Y = β0+ β1X1+ β2X2+ β3X3+ ε

From the equation, β0 is a constant, whereas β1 to β3 are coefficients of the independent variables. X1 X2 and X3 are the independent variables numbering of persons, rooms and bedrooms respectively, whereas epsilon ε is an error term. Additionally, the dependent variable Y in the equation represents public use microdata area code. Pegging the present discussion in the formula above, the model would be as follows.

Y = 231.130 + .114X1 – .121X2 +.128X3

This means that the public use micordata area code = 231.130 + (0.114 x numbering of persons) – (0.121 x rooms) +(0.128 x bedrooms).

References

Kothari, C. (2004). Research methodology, methods & techniques (2nd ed.).New Delhi: Wishwa Prakashan.

Saunders, M., Lewis, P. & Thornhill, A. (2007). Research Methods for Business Students. 4th edition. England: Prentice Hall.

Seber, A. F. G. and Lee, J. A. (2012) Linear Regression Analysis. 2nd Edition. Hoboken, New Jersey: John Wiley & Sons

Want help to write your Essay or Assignments? Click here

Statistics: Scatterplots Essay Paper

Scatterplots
Scatterplots

Statistics: Scatterplots

Question 1

Scatterplots have been ranked as one of the oldest and most common techniques for projecting high dimensional data to 2-dimensional form. Generally, these projections are arranged in a grid structure to aid the user in remembering the dimensions linked with each projection. Scatterplots have proven to be quite useful in determining the relationship between two variables.

Scatterplots have been linked with a number of benefits. For instance, the plots have been cited as one of the most important technique for studying non-linear pattern. Krzywinski & Altman (2014) state that it is easy to plot the diagram.  It is also useful in indicating the range of data flow, that is, the minimum and maximum value. Observation and reading of data in scatterplots is also straightforward.

Scatterplots are also important in studying large data quantities and make it easier to see the relationship between variables and their clustering effects. The use of scatterplots is an invaluable which is also useful in analyzing continuous data. However, this technique has a number of shortcomings such as being restricted generally to orthogonal views as well as challenges in projecting the relationship that exists in more two dimensions.

Question 2

When determining the appropriateness of the inferential statistical techniques, the researcher should first know if his/her data is arranged in a nested or crossed manner. If data is in a crossed manner, all study groups should have all intervention aspects. However, in a nested arrangement each study group will be subjected to a different variable. The correlation of variables can also be used in determining the appropriateness of a technique.

If the two variables have a linear or closely related association, the technique is said to be suitable. The number of assumptions that are employed when using a technique are also useful indicators. For instance, some techniques such as the t-test have several assumptions compared to the ANOVA technique, this implies that t-test has a large room for study errors unlike the ANOVA test.

Question 3

The Pearson product-moment correlation coefficient (PPMCC) is an analytical technique that is applied in indicating the strength of a linear association between two variables. This technique indicates a line of best fit through the data of two variables. It takes values ranging from +1 to -1 whereby a value of zero indicates that two variables of study do not have any association.

Values that are less than zero indicate that the two study variables have a negative association such that when one value increases the other one decreases. On the other hand, values that are greater than zero indicate that they have a positive association between them, that is, when there is an increase in one value, the other value increases as well. To determine whether the two variables have a linear relationship, they are fist plotted on a graph followed by a visual inspection of the shape of the graph.

A section of some scholars used to argue that PMCC can be used to indicate the gradient of a line. However, recent studies have dismissed this claim. For instance, Puth, Neuhäuser & Ruxton (2014) illustrate clearly that a coefficient of +1 does not mean that when one variable increases by one unit the other one increases by the same margin. A coefficient of +1 means that no variation exists between the plots and the line of best fit.

A number of assumptions must be put into consideration by analysts when they use the PPMCC. For instance, it is assumed that the outliers are either maintain at a minimum or removed completely. However, this is not the case in majority of studies. Another assumption is that the variables used should be distributed approximately normally and they must either be ratio or interval measurements.

References

Krzywinski, M., & Altman, N. (2014). Points of significance: visualizing samples with box plots. Nature methods, 11(2), 119-120.

Puth, M. T., Neuhäuser, M., & Ruxton, G. D. (2014). Effective use of Pearson’s product–moment correlation coefficient. Animal Behaviour, 93, 183-189.

Want help to write your Essay or Assignments? Click here