I have found this white paper about Clouds, Types of clouds, securing the cloud, I thought you may be interested to read it...
http://library.dzone.com/sites/all/files/whitepapers/Taking_Control_of_the_Cloud_0.pdf
I have found this white paper about Clouds, Types of clouds, securing the cloud, I thought you may be interested to read it...
http://library.dzone.com/sites/all/files/whitepapers/Taking_Control_of_the_Cloud_0.pdf
Regression vs. Retesting
You must retest fixes to ensure that issues have been resolved before development can progress.
So, retesting is the act of repeating a test to verify that a found defect has been correctly fixed.
Regression testing on the other hand is the act of repeating other tests in 'parallel' areas to ensure that the applied fix or a change of code has not introduced other errors or unexpected behavior.
For example, if an error is detected in a particular file handling routine then it might be corrected
by a simple change of code. If that code, however, is utilised in a number of different places
throughout the software, the effects of such a change could be difficult to anticipate. What appears to be a minor detail could affect a separate module of code elsewhere in the program. A bug fix could in fact be introducing bugs elsewhere.
You would be surprised to learn how common this actually is. In empirical studies it has been estimated that up to 50% of bug fixes actually introduce additional errors in the code. Given this,
it's a wonder that any software project makes its delivery on time.
Better QA processes will reduce this ratio but will never eliminate it. Programmers risk
introducing casual errors every time they place their hands on the keyboard. An inadvertent slip of a key that replaces a full stop with a comma might not be detected for weeks but could have
serious repercussions.
Regression testing attempts to mitigate this problem by assessing the ‘area of impact’ affected by a change or a bug fix to see if it has unintended consequences. It verifies known good behavior after a change.
What is hadoop?
• Open source software for reliable, scalable, and distributed computing.
• Flexible infrastructure for large scale computation and data processing on a network of commodity hardware.
• The Linux of distributed processing.
Why hadoop?
• Very large distributed file system.
• The data is distributed across data nodes .
• Reliability and availability.
• Files are replicated to handle hardware failure.
• Detects failures and recovers from them.
• Ability to run on cheap hardware.
• Open source flexibility.
• Runs on heterogeneous OS.
• Scalability.
• The number of nodes in a cluster is not constant.
• Parallel processing through MapReduce.
Main components:
• HDFS(Hadoop file system) for storing.
• The Map-Reduce programming model for processing.
Hadoop Distributed File System (HDFS)
• A distributed file system based on GFS, as its shared filesystem.
• Distributed across data servers.
• Data files partitioned into large chunks (64MB), replicated on multiple data nodes.
• NameNode stores metadata information (block locations, directory structure).
Map-Reduce:
• Framework for distributed processing of large data sets.
Copyright © 2013 Tech-Buzz.