Month: January 2014
SharePoint farm solution
Farm solutions, which are hosted in the IIS worker process (W3WP.exe), run code that can affect the whole farm. When you debug a SharePoint project whose Sandboxed Solution property is set to “farm solution,” the system’s IIS application pool recycles before SharePoint retracts or deploys the feature so as to release any files locked by the IIS worker process. Only the IIS application pool serving the SharePoint project’s site URL is recycled. SharePoint 2013 uses the traditional software licensing and installation. Known as an on-premises solution, customers can install and configure SharePoint to deploy a farm on physical computers or in a virtual environment. Customers also have the option of deploying and operating a farm in their own data center or on an infrastructure provided by a hosting service
SharePoint Farm 2013 Azure Deployment
First of all must be mentioned that this analysis is a brief and global one. The timescale to do a thorough analysis of the needed farm solution and the related pricing model would take more time than given. Microsoft sources are used to come up with some rough indications and all sources are properly referenced.Small server farm A small server farm consists of at least two Web servers and a database server. One of the Web servers hosts the Central Administration site and services and theother handles additional tasks, such as handling content requests. This farm can be scaled out to three tiers by using a dedicated application server. Medium server farm
A medium server farm usually has two or more Web servers, two application servers, and at least two database servers. Large server farm A large server farm is the result of scaling out a medium farm to meet capacity and performance requirements or in preparation for implementing a SharePoint 2013 solution. A three-tier topology typically uses dedicated servers on each tier and servers are grouped according to their role
Farm topology examples
The following examples show how SharePoint can be deployed on one or more servers using tiered topologies and server roles to implement a farm design that meets specific goals and objectives.
There are scenarios where it makes sense to install SharePoint on single server. A single server farm is typically used to demonstrate SharePoint, to conduct a cursory evaluation of SharePoint features, or to compare SharePoint 2013 with previous releases of the product.
In two-tier farm, the database server is installed on one server and all the other roles are installed on the second server. This level of separation is the minimum we recommend for deploying SharePoint on Windows Azure. This environment is well suited for in-depth product evaluation,
The three-tier topology shown in the following illustration consists of two front-end Web servers, an application server, and a database server. This model provides the foundation for deploying a farm that can scale out in response to increased workload demands or an expanding user base. It also provides the framework for using redundant servers to increase farm capacity and at the same time increase farm availability, which is shown in Figure 3. It is worth noting that in order to provide a highly available service you will need at least 2 servers in each role.
Operate and maintain SharePoint farm on Windows Azure
Operating a SharePoint farm running in Windows Azure is no different than operating a SharePoint farm anywhere else. Despite the power and agility Windows Azure brings you, at the end of the day you still have a Windows Server, SharePoint and the dependencies to manage. On premises you may use a combination of Windows Server tools and the SharePoint administration tools, or manage the whole farm via System Center. Regardless, the same tools and procedures can be used when the farm is hosted in Windows Azure.
For more information see the monitoring guide at Plan for monitoring in SharePoint 2013 http://technet.microsoft.com/en-us/library/jj219701.aspx
Patching and updating
Again, patching and updating the OS, SharePoint and the dependencies is not different in Windows Azure than anywhere else. You are still responsible for OS patching etc. and you are still in full control of how and when that should happen. Patching and updating a SharePoint 2013 farm can be quiet involved depending upon the topology. Please refer to http://technet.microsoft.com/en-us/library/ff806329.aspx for more information.
One area in which Windows Azure can help is when testing major updates. Since you can create resources on-demand, you could spin up a duplicate environment within Windows Azure and test your update or patching strategy and methodology.
Backup and Recovery
Backing up and recovery of SharePoint farms in Windows Azure is again is very similar. One thing to consider is that you should suffer no significant down-time due to hardware failure. Since Windows Azure will auto repair and redeploy your virtual machine there is no action to take on your part. That said, you do need to consider the fact that you will get “hardware” downtime that would be automatically repaired. It is good to ensure that any customizations you perform or applications you deploy can handle this automatic recovery. The second key point here is that Windows Azure makes it very easy to deploy more virtual machines making it possible to create a highly available farm as discussed previously.
E-commerce Platform is a software technology solution that allows you to build storefronts Facebook Website In Store Mobile app Call Center store eCommerce Platform.E-Commerce is an absolute requirement for success in today’s fast-paced global economy. Unique possibilities exist to widen markets, reduce costs, increase efficiency and offer customers seamless access to information. To capitalize on these opportunities, organizations globally including many of your competitors and suppliers are turning to e-Commerce as a competitive necessity.
Features of E-commerce development
- Easy to Adapt and Flexible
- Scalable and Secure
- Open Platforms
Benefits of e-Commerce
It contributes through boosting efficiency from consistent, automated, accurate information available in real time.
- Improved customer relations.
- A comparative reduction in the cost of operations along with lesser complexity.
- Decrease in operation costs resulting from streamlined infrastructure of E-Commerce
The fact that people want it is probably the biggest lure for companies to jump onto the bandwagon, but there are other factors, making e-commerce a good sense solution.
– Lower transaction costs. If the site is implemented well, the web can significantly lower order taking costs and customer service costs after the sale by automated processes.
– Variety for shoppers: It gives people the opportunity to shop in different ways.
– The ability to build an order over several days
– The ability to configure products and see actual prices
– The ability to compare prices between multiple vendors
– The ability to search large catalogues easily
NEW DELHI: Tech giant Microsoft has said that public sector Bank of India has deployed itsSharePoint solution to automate the loan and application deposit process in its over 4,100 branches and 50 zonal offices across India.
The automation of 15 different loan application forms have reduced the turn-around time for filling up applications by 50%, improved customer satisfaction and increased operational efficiency, Microsoft said in a release.
Bank of India is an early adopter of Microsoft SharePoint and has set this as a benchmark for other nationalized banks to increase efficiency and collaboration, especially for customers who prefer online transactions and access their bank accounts over the internet, it added.
Microsoft is making good on its commitment to add more e-mail and SharePoint integration to its Yammer enterprise social-networking product.
- Comprehensive collaboration. The network unites response agents in the field with claims processors at headquarters, who can tap in on any device.
- A more effective social intranet. Yammer integrates with Nationwide’s key applications, including SharePoint.
- Improved productivity. Better information sharing and the crowdsourcing of ideas means faster responses to business and customer demands.
- A stronger corporate culture. The network helped transform a widespread employee base into a more tightly knit workforce focused on customer satisfaction
The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing.
The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures.Hadoop has moved far beyond its beginnings in web indexing and is now used in many industries for a huge variety of tasks that all share the common theme of lots of variety, volume and velocity of data – both structured and unstructured. It is now widely used across industries, including finance, media, entertainment,government, healthcare, information services, retail, and other industries with Big Data requirements but the limitations of the original storage infrastructure remain.
Apache Hadoop includes a Distributed File System (HDFS), which breaks up input data and stores data on the compute nodes. This makes it possible for data to be processed in parallel using all of the machines in the cluster. The Apache Hadoop Distributed File System is written in Java and runs on different operating systems.Apache Hadoop is 100% open source, and pioneered a fundamentally new way of storing and processing data. Instead of relying on expensive, proprietary hardware and different systems to store and process data, Hadoop enables distributed parallel processing of huge amounts of data across inexpensive, industry-standard servers that both store and process the data, and can scale without limits. With Hadoop, no data is too big. And in today’s hyper-connected world where more and more data is being created every day, Hadoop’s breakthrough advantages mean that businesses and organizations can now find value in data that was recently considered useless.
Hadoop Different from Past Techniques?
- Hadoop can handle data in a very fluid way
- Hadoop has a simplified programming model
- Hadoop is easy to administer
- Hadoop is agile
Benefits of Apache Hadoop?
- cost effective.
Overview of Hadoop