Want help to write your Essay or Assignments? Click here

Legislative Critique of the Maritime Transportation Security Act 2002 and the S.A.F.E. Port Act of 2006.

There are two significant pieces of legislation that impact maritime security in the United States. These include the Maritime Transportation Security Act 2002 and the S.A.F.E. Port Act of 2006.

For this assignment, you are to conduct a graduate level Legislative Critique of these two pieces of legislation. To do this critique perform the following steps:

. Obtain a copy of the piece of legislation to be examined. (MTSA and S.A.F.E. Port Act of 2006)

2. Collect and analyze news media reports on the legislation.

3. Collect and analyze any other critiques of the legislation. (i.e. Think tanks, academics, etc)

4. Collect and analyze available witness testimony with regard to the legislation.

To compose a legislative critique students should compose an analysis based upon the following points:

• Provide a brief legislative history of the Bills. (No more that 500 words – USE WORD ECONOMY)!

• Determine if the Bills focus was introduced in the past under a different name and/or format. If so, evaluate why the legislation was discarded.

• Summarize the main points of each bill in your own words (but supported by citation as appropriate)

• Evaluate and categorize the strengths and weaknesses of the individual legislation, then compared against each other.

• Conclude with a final assessment of both pieces of legislation including SPECIFIC recommendations that could be considered to enhance the security framework that has been established.

We can write this or a similar paper for you! Simply fill the order form!

Sampling Distributions and Confidence Intervals

ORDER INSTRUCTIONS:

1. What is the central limit theorem? What role does it play in the establishment of the sampling distribution of the sample mean? 2. Contrast normal distributions with sampling distributions.

In statistical inference, one wishes to estimate population parameters using observed sample data.

A confidence interval gives an estimated range of values which is likely to include an unknown population parameter, the estimated range being calculated from a given set of sample data. (Definition taken from Valerie J. Easton and John H. McColl’s Statistics Glossary v1.1)

The common notation for the parameter in question is . Often, this parameter is the population mean , which is estimated through the sample mean .

Below is a partial answer to the above homework questions by one of our writers. If you are interested in a custom non plagiarized top quality answer, click order now to place your order.

Sampling Distributions and Confidence Intervals Essay

The central limit theorem

The central limit theorem refers to a theory in probability statistics where the approach is that the distributions of a normalized….

The role of the Central Limit Theorem in the establishment of the sampling distributions of the sample mean

A major role of the central Limit theorem in terms of establishing sampling distributions is the use of inference from the available….

Contrasting normal distributions with sampling distributions

Normal distributions refer to a normal continuous probability distribution that when….

We can write this or a similar paper for you! Simply fill the order form!

Want help to write your Essay or Assignments? Click here

Bare.nuclei, bare nuclei

Bl.cromatin, Bland cromatin

Normal.nucleoli, normal nucleoli

Mitoses

Class (this is the target variable)

The goal is to predict values of Class.

Variables 2 through 10 are coded as factors: for some of the work you may want to convert them to numerical values and standardize them.

Want help to write your Essay or Assignments? Click here

In your analysis

Examine the data for missing values, and explain how you decided to deal with them

Examine the distributions of variables 2 through 10 (boxplots or histograms, or both) for shape and outliers. These variables are recorded so that all are scaled to a range of 0 to 10, but their distributions may not be (probably are not) the same, hence the need for investigation. If you want to create histograms, note that the hist() function in R requires a vector of values, so you will need to do something like hist(as.numeric(BreastCancer$cell.size)) to get the histogram to work

Prepare the data for use in a neural network

Fit a single hidden layer neural network to the data and assess its performance. (Do the usual training/testing process). Try to tune the model to select an ‘optimal’ value of hidden nodes

Use network averaging to model the data, and assess its performance

We can write this or a similar paper for you! Simply fill the order form!

Want help to write your Essay or Assignments? Click here

Section One: Calculation of Full-Time Equivalents

Personnel Budget Case Study

Question One

There are a number of items that are required in order to ensure that a budget is effectively prepared. According to lectures of this course, several items that are required include: patient acuity, patient days, length of a patient stay, number of the vacancies that are anticipated in each level, staffs’ educational needs in the in the year that follows, costs associated with staff benefits, as well as non-productive staff time that is anticipated. The background information/data about 1 West unit is as follows:

Patient Data: Average Daily Census (30); Unit Capacity (32); Average HPPD (8.8); and Total Care Hours (96,360)

Staff Data: Productive hours/employee/year (1,780); Nonproductive hours/employee/year (300); and Total Hours/employee/year (2,080)

Skill Mix: RNs (80%); LVNs (10%); and Nurse Aides (10%)

Want help to write your Essay or Assignments? Click here

Question Two

Calculating the number of the productive FTEs to be needed:

The initial step would be to consider the subsequent year’s workload, figuring in mind that the anticipated number of patient days will be 10,950 [obtained by multiplying average daily census (30) by 365 days].

Step 1:

Calculating Workload = HPPD x Number of patient days = 8.8 x 10,950 = 96,360

Step 2:

Calculating Productive time = Subtracting total nonproductive hours from total FTE hours; 2,080 – 300 = 1,780 hours

Step 3:

Calculating FTEs = Workload divided by productive time: 96,360/1,780 = 54.3 productive FTEs needed to staff the unit.

Want help to write your Essay or Assignments? Click here

Question Three

Determining the number of RNs, LVNs, as well as nurse aides that will be required to ensure that the unit is well staffed based on the assumption that there will be a 12-hours working shifts for the staff. In other words, this requires the calculation of the number of persons-shifts to be need during periods of 24-hour shifts.

Since the staffs are shared 50% for day shift and 50% for night shift in the following mix RNs (80%); LVNs (10%); and Nurse Aides (10%), the numbers are as follows:

In a 12-hour shift:

RNs = 80/100 * 54.3 = 43.44

LVNs = 10/100 * 54.3 = 5.43

Nurse aides = 10/100 * 54.3 = 5.43

Want help to write your Essay or Assignments? Click here

Question Four

Assigning the staff by shift and by type considering that:

Day Shift 50%

Night Shift 50%

Then,

RNs = 80/100 * 54.3 = 43.44 each shift

LVNs = 10/100 * 54.3 = 5.43 each shift

Nurse aides = 10/100 * 54.3 = 5.43 each shift

Want help to write your Essay or Assignments? Click here

Question Five

Coverting staff positions to full-time Equivalents or FTE positions involves converting to 24/7 as follows:

Total FTEs = 54.3 x 1.4 = 76.02 FTEs needed to staff the unit 24/7, which is equivalent to 76.02 FTE positions

This will help convert staff positions to FTE positions needed for a 24/7 shift as follows:

RNs = 80/100 * 76.02 = 60.8

LVNs = 10/100 * 76.02 = 7.6

Nurse aides = 10/100 * 76.02 = 7.6

Want help to write your Essay or Assignments? Click here

Section Two: Variance Analysis

Variance Analysis Case Study

Comparing the original budget to the flexible budget reveals that, the original one had no significant unfavorable variances compared to the flexible one, which is characterized by unfavorable variances not only in terms of volume but also in terms of price and quantity. For instance, a volume variance is experienced when the actual volume is higher or lower than the budgeted volume and may be expressed in terms of FTEs or patient days. To determine volume variance, the calculation is as follows:

(Budgeted Volume – Actual Unit Volume)(Budgeted Rate) = Volume Variance

In order to get the budgeted rate, the following formula is used:

Want help to write your Essay or Assignments? Click here

A comparison between the flexible budget and actual budget shows that, there are unfavorable variations in terms of price and quantity. This is because employees’ average hourly rate has increased from $40.00 to $45.00, whereas hours per care per patient have increased from 5.0 to 5.6. These two changes are indicative of an unfavorable variation. Price variance or unit cost variance determination can be calculated using the equation shown below:

(Budgeted Unit Price – Actual Unit Cost)(Actual Volume) = Unit Cost Variance

Want help to write your Essay or Assignments? Click here

There are several factors that led to the differences in variances including both external and internal factors. In particular, the internal factors include changes in staff efficiency which has led to an increase in the hours per care per patient from 5.0 to 5.6. Changes in technology as well as nature of surgeries may have also be other internal factors that have led to these variations.

This is because all of these factors can collaboratively combine to delay the rate at which patients are operated eventually increasing the overall average time required to provide care to each patient in hours per visit. On the other hand, external factors that may have caused the variations include type of staff available, census changes as well as price changes, all of which can be attributed to an increase in the hours per care per patient from 5.0 to 5.6, number of visits from 340 to 400 as well as employees’ average hourly rate from $40.00 to $45.00. The unfavorable variations occurred because the above discussed factors combined to negatively impact the budgeted figures.

Want help to write your Essay or Assignments? Click here

Statistics in Prediction of Outcome

Introduction

Statistic can be used to predict outcome. This is because statistic encompasses the use of various techniques that enable a person in prediction of outcome. To emphasize this fact, we carried out a project. Our project was about analyzing data to extend where we answer the research question. The topic was chosen because there is need to understand the relationship between statistic and outcome. We wanted to come up with proper understanding of the reliability of statistic in prediction of outcome. The research question was ‘can statistic predict income?’

Saint Leo University has excellence as one of the core values. The excellence that the University dreams of is about creating character that enables each one of the students to end up responsible people in the society. This is brought about by team work, dedication and completion of assigned tasks. Therefore, completing this project reflected what the value of excellence at Saint Leo University advocates.

Data

The data that we used is secondary. The data used in this report was obtained from the website of Halliburton Limited, a multinational company operating in the oil fields service providers industry.

In this project, we used dependent and independent variables. Dependent variables are the ones that a researcher is always interested in. On the other hand, the independent variables are the ones that affect the dependent variables.

Want help to write your Essay or Assignments? Click here

Dependent and independent variables

In our project, the dependent variables consist of operating income for three years. On the other hand, the independent variables consist of sales values for three years. According to Halliburton (2014), the sales and operating income values were as below.

Year

Sales (x)

Profit(y)

2012

28,503.00

4,159.00

2013

29,402.00

3,138.00

2014

32,870.00

5,097.00

Descriptive Statistics Table

Statistic

Sales (x)

Profit(y)

Mean

30,258.33

4,131.33

Median

29,402.00

4,159.00

Min

28,503.00

3,138.00

Max

32,870.00

5,097.00

Standard Deviation

1882.84

799.998

Skewness

1.36

0.10

Statistic Test

The testing that we carried out is Pearson coefficient of correlation. This test is quite useful in understanding the relationship of variables. It is very useful in understanding how strong the relationship between variables is.

Year

Sales (x)

Operating Income (y)

(xy)

(x^2)

(y^2)

2014

32,870.00

5,097.00

167,538,390.00

1,080,436,900.00

25,979,409.00

2013

29,402.00

3,138.00

92,263,476.00

864,477,604.00

9,847,044.00

2012

28,503.00

4,159.00

118,543,977.00

812,421,009.00

17,297,281.00

Total

90,775.00

12,394.00

378,345,843.00

2,757,335,513.00

53,123,734.00

The formulae of Pearson Coefficient of correlation is as shown below

N

3.00

N∑xy

1,135,037,529.00

∑x∑y

1,125,065,350.00

N∑x^2)

8,272,006,539.00

∑x^2)

2,757,335,513.00

N∑y^2)

159,371,202.00

∑y^2)

53,123,734.00

Coefficient of correlation (r) = 1,135,037,529.00-1,125,065,350.00/ Square root of (8,272,006,539.00-2,757,335,513.00)( 159,371,202.00- 53,123,734.00)

After conducting Pearson coefficient of correlation, the result was a positive correlation of the variables. This means that when sales increase, the operating income increases too. This shows that we can predict the outcome of the year 2015.

Conclusion

From the project, it is evident that statistic can be used to predict outcome. Therefore, the research question has been addressed well. From the analysis, it has become evident that the use of statistic is faced with certain challenges. Firstly, reliability of some data sources is questionable. Additionally, the fact that data is not readily available is another challenge. There has also been an important observation that the choice of variables should be done properly to avoid confusion of the tests carried out. From the project, it has come out clearly that when there is positive correlation of variables, the movement of change in any of them is positive.

Inferential Statistics and Findings: Generating a Hypothesis

The aim of the paper is to generate a hypothesis with the use of research questions and two variables- dependent variable: sales/revenue and independent variable: interest rates. In addition, the paper uses a statistical technique for testing the hypothesis at 95% confidence interval while interpreting the results. XYZ Corporation is a US-based company that majors in home improvement and construction.

Moreover, the company operates in various big-box format stores across the United States. Nevertheless, XYZ Corporation wants to determine the effects of interest rates and sales on its future profitability, therefore, the company approached a research team that suggested the below research question.

Research question

Is there a relationship between federal (Feds) directed interest rate (ID) and sales /revenue (DV) on future profit of XYZ Corporation?

Hypothesis H0 = XYZ Corporation’s future profits are not determined by interest rates and sales/revenue H1 = XYZ Corporation’s future profits are determined by interest rates and sales/revenue

Hypothesis tests with a 95% confidence level

The veracity of the null hypothesis depends on data as well as the objective of significance tests to estimate the likelihood, hence the P-value, which underscores that the occurrence is an issue of chance. In this case, if the Pearson Correlation is less than 5% it means that the null hypothesis is rejected.

However, if the P-value is greater than 5% then it means the statistical test fails to reject the null hypothesis (Gupta, 2012). Data was indiscriminately selected from 382 observations with respect to interest rates and sales of XYZ Corporation.

Interpretationof the results and findings

Statistical analysis demonstrates that this is a normal distribution, with a mean of 10.97644 and standard deviation of 10.14628 for sales (Table 1). In this case, sales were observed between $0.00 and $89.00. Changes in the federal interest rates could help determine sales of XYZ Corporation.

In addition, the mean and standard deviation for interest rates were observed at 9.695538 and 7.836266 respectively (Table 2). High-interest rates are highly likely to inhibit any profit making from lack of sales. While low-interest rates by the federal government may lead to increased sales.

Again, it demonstrated a 95% confidence interval (sales:1.020716, interest rates: 0.789369) this implies that XYZ Corporation can be about 95 percent confident that sales and interest rates will lead to an increased future a profit. Because the confidence interval of sales corresponds to the value of alpha, which is 1 indicates that there is no significant correlation between sales and future profit of XYZ Corporation.

However, the results are empirically reliable. When it comes to interest rates the confidence interval is less than 1, statistically the test fails to refute the null hypothesis, an aspect that demonstrates statistical significance of the test (Monette, Sullivan & DeJong, 2011).

Gupta, S. K. (2012). The relevance of confidence interval and P-value in inferential. Indian Journal of Pharmacology. Retrieved on June 9, 2016, from statistics http://www.ncbi.nlm.nih.gov/pmc/articles/PMC3271529/

Monette, D. R., Sullivan, T. J., & DeJong, C. R. (2011). Applied Social Research: A Tool for the Human Services (8th ed.). Belmont, CA: Brooks/Cole.

Want help to write your Essay or Assignments? Click here

This report is an analysis of the data generated using SPSS and presented using charts and tables. The report firstly presents the results of selected descriptive statistical analyses. Subsequently, the report summarises the numerical results with descriptive statistics analysis tables or graphs, including the interpretation of these tables and graphs. The fourth section or the report is a presentation of the data regarding numerical results of the inferential statistics. This is followed by a discussion of the same, before a summative conclusion is presented in the last section.

Selected descriptive statistics

Descriptive statistics refers to the kinds of data that analysts and researchers use in presenting the characteristics of the sample used in a study. According to Kothari (2004), they are used in checking whether the variables that the researcher has chosen to use violate any assumptions that the researcher might have made, which might be consequential to the findings. Another important function of descriptive statistics used in this section is that they help to answer the core research questions.

In the present study, the descriptive statistics selected are for public use micro data area code (PUMA), house weight (WHTP), state code, (ST), numbering of persons (NP), rooms (RMS), bedrooms, (BDS), and household income (HINCP). The data retrieved was as presented in table 1 below

Table 1: PUMA, ST, BDS, RMS, mean, median, and standard deviation

RMS

BDS

ST

PUMA

N

Valid

4911

4911

4911

4911

Missing

0

0

0

0

Mean

4.87

2.61

15.00

248.05

Median

5.00

3.00

15.00

302.00

Std. Deviation

1.933

1.197

.000

81.573

Minimum

1

0

15

100

Maximum

9

5

15

307

Table 2: RMS, BDS, ST, and PUMA, frequency table

PUMA

Frequency

Percent

Valid Percent

Cumulative Percent

Valid

100

951

19.4

19.4

19.4

200

782

15.9

15.9

35.3

301

407

8.3

8.3

43.6

302

412

8.4

8.4

52.0

303

425

8.7

8.7

60.6

304

536

10.9

10.9

71.5

305

365

7.4

7.4

79.0

306

456

9.3

9.3

88.3

307

577

11.7

11.7

100.0

Total

4911

100.0

100.0

NP

Frequency

Percent

Valid Percent

Cumulative Percent

Valid

0

452

9.2

9.2

9.2

1

970

19.8

19.8

29.0

2

1491

30.4

30.4

59.3

3

711

14.5

14.5

73.8

4

619

12.6

12.6

86.4

5

318

6.5

6.5

92.9

6

161

3.3

3.3

96.2

7

94

1.9

1.9

98.1

8

30

.6

.6

98.7

9

20

.4

.4

99.1

10

18

.4

.4

99.5

11

7

.1

.1

99.6

12

7

.1

.1

99.7

13

5

.1

.1

99.8

15

1

.0

.0

99.9

16

2

.0

.0

99.9

17

1

.0

.0

99.9

19

2

.0

.0

100.0

20

2

.0

.0

100.0

Total

4911

100.0

100.0

RMS

Frequency

Percent

Valid Percent

Cumulative Percent

Valid

1

185

3.8

3.8

3.8

2

345

7.0

7.0

10.8

3

677

13.8

13.8

24.6

4

896

18.2

18.2

42.8

5

1110

22.6

22.6

65.4

6

768

15.6

15.6

81.1

7

438

8.9

8.9

90.0

8

234

4.8

4.8

94.7

9

258

5.3

5.3

100.0

Total

4911

100.0

100.0

BDS

Frequency

Percent

Valid Percent

Cumulative Percent

Valid

0

211

4.3

4.3

4.3

1

683

13.9

13.9

18.2

2

1208

24.6

24.6

42.8

3

1810

36.9

36.9

79.7

4

688

14.0

14.0

93.7

5

311

6.3

6.3

100.0

Total

4911

100.0

100.0

From the data in table 1 above, a number of observations are blatant and clear. The first is that the means of RMS, BDS, ST and PUMA are 4.87, 2.61, 15, and 248.05 respectively. For rooms, the number of rooms, the median score was 5, where the scores varied from 1 to 9. This means that the majority of respondents have about 5 rooms.

When it comes to the number of bedrooms, the median score was 3, whereas the mean was 2.61, this shows that the majority of respondents have 3 rooms. The state code was 15 for all respondents whereas the mean for public use of micro data area code was 248.05. The mean was 302, whereas the minimum and maximum scores were 100 and 307 respectively.

From table 2, a number of assertions can also be made, and the first is about PUMA. From the table, the evidence shows that for public use of micro data area code, 19.4% of the respondents scored category 100, which made it the highest selected category, whereas 15.9% of the respondents checked 200, making it the second most selected category. Comparatively, 301 was the least selected category at 8.7%.

Additionally, for number of bedrooms, a majority of the respondents said that they had three bedrooms in their houses, and this represented 3.9% of all responses, closely followed by those with two bedrooms at 24.6%. At the same time, the number of people living in houses with no bedrooms or five bedrooms was the least with a score of 4.3% and 6.3% respectively.

This data is in line with the data about rooms, which shows that 22% of respondents stay in a five-roomed apartment, followed by 18% and 15%, who stay in four and five roomed houses respectively. Because of the number of rooms and bedrooms in their houses, it is plausible to conclude that a majority of the respondents stay with other people or expect other people to visit often, which are why they have extra rooms in the house, as well as extra bedrooms in the house.

Additionally, from the data, it is obvious that a majority of the people are in the middle between the rich and the poor, as those who stay in studio apartments are as marginal as those who stay in luxury apartments that can contain at least five bedrooms. .

Selected inferential statistical analyses

Inferential statistics refer to the data analysis methods where the researcher or analyst uses a given set of data to determine whether there is a link between given variables being studied. By using inferential statistics, the researcher can tell whether the relationship that seems to exist between variables is a fact, or whether it is not a fact. According to Kothari (2004), a number of measures and techniques can be used to accomplish inferential statistics. The two types of inferential statistics used in this report are correlation and regression analyses.

Correlation was conducted using the Pearson correlation analysis. Pearson correlation analysis is employed to measure the linear relationship between two or more variables. The value of Pearson correlation ranges between -1 and +1, with -1 indicating negative correlation, 0 indicating no correlation and +1 indicating positive correlation between the variables. Besides, the closer the value is to +1, the stronger the relationship between the variables (Saunders, Lewis & Thornhill, 2007). For this study, the data is as shown below.

According to table 4-20, Sig. (2-tailed) =0.000, and all the four variables have a significant correlation at the 0.01 significant level. Pearson correlation between PUMA and NP is .110, whereas the relation between PUMA and BDS and RMS is .042 and .067 respectively. This shows that there is a weak but positive relationship between PUMA and all the independent variables, although the weakest relationship is that between PUMA and BDS.

Table 3: Correlations

PUMA

NP

BDS

RMS

PUMA

Pearson Correlation

1

.110(**)

.042(**)

.067(**)

Sig. (2-tailed)

.000

.003

.000

N

4911

4911

4911

4911

NP

Pearson Correlation

.110(**)

1

.447(**)

.396(**)

Sig. (2-tailed)

.000

.000

.000

N

4911

4911

4911

4911

BDS

Pearson Correlation

.042(**)

.447(**)

1

.878(**)

Sig. (2-tailed)

.003

.000

.000

N

4911

4911

4911

4911

RMS

Pearson Correlation

.067(**)

.396(**)

.878(**)

1

Sig. (2-tailed)

.000

.000

.000

N

4911

4911

4911

4911

** Correlation is significant at the 0.01 level (2-tailed).

Regression analysis helps estimate and investigate the association between variables. R Square is used to show the degree of relationship between the dependent and independent variables. R Square value ranges between 0 and 1, and the closer the value is to 1, the stronger the relationship between the variables further indicating the greater degree to which variation in independent variable explains the variation in dependent variable (Seber and Lee, 2012).

Based on the model summary table 4-21, R stand for the correlation coefficient and it depicts the association between dependent variable and independent variables. It is evident that a positive relationship exists between the dependent variable and independent variables as shown by R value (0.126).

However, the relationship is a very weak one. Besides, it can be seen that the variation in the three independent variables (RMS, BDS and NP) explain 1.6% variation of PUMA as represented by the value of R Square. Therefore, it means that other factors that are not studied on in this study contribute 98.4% of the PUMA programs. This means that the other factors are very important and thus need to be put into account in any effort to enhance PUMA. Additionally, this research therefore identifies the three independent variable studied on in this research as the non-critical determinants of PUMA boundaries.

Table 4: regression analysis results

Model Summary

Model

R

R Square

Adjusted R Square

Std. Error of the Estimate

1

.126(a)

.016

.015

80.945

a Predictors: (Constant), RMS, NP, BDS

Further, this research established through the analysis f variance that the significant value is 0.00, which is less than 0.01, therefore the model is statistically significant in foretelling how NP, RMS, and BDS can influence PUMA groupings. The F critical value at the 0.01 level of significant was 26.501. Given that F calculated is greater than the F critical value of 26.501, then it means that the overall model was significant (Seber and Lee, 2012).

ANOVA(b)

Model

Sum of Squares

df

Mean Square

F

Sig.

1

Regression

520911.120

3

173637.040

26.501

.000(a)

Residual

32151092.845

4907

6552.087

Total

32672003.965

4910

a Predictors: (Constant), RMS, NP, BDS

b Dependent Variable: PUMA

At the same time, the beta coefficients also gives significant inferential information. According to the regression coefficients presented in table 4-23, this research found that when all independent variables (the number of persons (NP), number of rooms (RMS), and the number of bedrooms (BDS)) are kept constant at zero, the level of public use micro data area code (PUMA) will be at 231.13. A 1% change in number of persons will lead to an 11.4% increase in PUMA, whereas a one percent change in BDS will lead to a 12.1% changes in PUMA.

Comparatively, a one percent change in RMS will lead to a 12.8 percent change in PUMA. This leads to the conclusion that of the three variavles, RMS leads to the largest impact in PUMA when the three independent variables are pitted together. Further the statistical significance of each independent variable was tested at the 0.01 level of significance of the p-values.

Coefficients(a)

Model

Unstandardized Coefficients

Standardized Coefficients

t

Sig.

B

Std. Error

Beta

1

(Constant)

231.130

3.161

73.128

.000

NP

4.700

.654

.114

7.181

.000

BDS

-8.222

2.068

-.121

-3.977

.000

RMS

5.384

1.248

.128

4.315

.000

a Dependent Variable: PUMA

In general form, it can be said that the equation used to determine the link between Public use microdata area code, numbering of persons, rooms and bedrooms is of the form:

Y = β_{0}+ β_{1}X_{1}+ β_{2}X_{2}+ β_{3}X_{3}+ ε

From the equation, β_{0} is a constant, whereas β_{1 }to β_{3 }are coefficients of the independent variables. X1 X2 and X3 are the independent variables numbering of persons, rooms and bedrooms respectively, whereas epsilon ε is an error term. Additionally, the dependent variable Y in the equation represents public use microdata area code. Pegging the present discussion in the formula above, the model would be as follows.

Y = 231.130 + .114X1 – .121X2 +.128X3

This means that the public use micordata area code = 231.130 + (0.114 x numbering of persons) – (0.121 x rooms) +(0.128 x bedrooms).

References

Kothari, C. (2004). Research methodology, methods & techniques (2nd ed.).New Delhi: Wishwa Prakashan.

Scatterplots have been ranked as one of the oldest and most common techniques for projecting high dimensional data to 2-dimensional form. Generally, these projections are arranged in a grid structure to aid the user in remembering the dimensions linked with each projection. Scatterplots have proven to be quite useful in determining the relationship between two variables.

Scatterplots have been linked with a number of benefits. For instance, the plots have been cited as one of the most important technique for studying non-linear pattern. Krzywinski & Altman (2014) state that it is easy to plot the diagram. It is also useful in indicating the range of data flow, that is, the minimum and maximum value. Observation and reading of data in scatterplots is also straightforward.

Scatterplots are also important in studying large data quantities and make it easier to see the relationship between variables and their clustering effects. The use of scatterplots is an invaluable which is also useful in analyzing continuous data. However, this technique has a number of shortcomings such as being restricted generally to orthogonal views as well as challenges in projecting the relationship that exists in more two dimensions.

Question 2

When determining the appropriateness of the inferential statistical techniques, the researcher should first know if his/her data is arranged in a nested or crossed manner. If data is in a crossed manner, all study groups should have all intervention aspects. However, in a nested arrangement each study group will be subjected to a different variable. The correlation of variables can also be used in determining the appropriateness of a technique.

If the two variables have a linear or closely related association, the technique is said to be suitable. The number of assumptions that are employed when using a technique are also useful indicators. For instance, some techniques such as the t-test have several assumptions compared to the ANOVA technique, this implies that t-test has a large room for study errors unlike the ANOVA test.

Question 3

The Pearson product-moment correlation coefficient (PPMCC) is an analytical technique that is applied in indicating the strength of a linear association between two variables. This technique indicates a line of best fit through the data of two variables. It takes values ranging from +1 to -1 whereby a value of zero indicates that two variables of study do not have any association.

Values that are less than zero indicate that the two study variables have a negative association such that when one value increases the other one decreases. On the other hand, values that are greater than zero indicate that they have a positive association between them, that is, when there is an increase in one value, the other value increases as well. To determine whether the two variables have a linear relationship, they are fist plotted on a graph followed by a visual inspection of the shape of the graph.

A section of some scholars used to argue that PMCC can be used to indicate the gradient of a line. However, recent studies have dismissed this claim. For instance, Puth, Neuhäuser & Ruxton (2014) illustrate clearly that a coefficient of +1 does not mean that when one variable increases by one unit the other one increases by the same margin. A coefficient of +1 means that no variation exists between the plots and the line of best fit.

A number of assumptions must be put into consideration by analysts when they use the PPMCC. For instance, it is assumed that the outliers are either maintain at a minimum or removed completely. However, this is not the case in majority of studies. Another assumption is that the variables used should be distributed approximately normally and they must either be ratio or interval measurements.