02-03-2013, 11:22 AM
Equity Trading Portfolio Manager
Equity Trading.doc (Size: 3.07 MB / Downloads: 55)
Abstract
In the Equity Share Portfolio Manager, the main objective of developing online share Trading system .Net Project is to provide effective trading tool over the internet. This is helps users to buy and sell company shares online. This system works with the new user registration by clicking on create new user link on homepage, check the different company trading share values, buy selected company’s number of shares, sell shares which already bought earlier, find out he price alerts, know about the price alerts, and user account balance details.
The Equity Share Portfolio Manager does not settle with a credit supplier of the customer until the item selected by the customer is picked from inventory but before it is delivered. Therefore, the customer can go online and make changes to the order. In addition, available service windows are presented to the customer as a function of customer selected order and service types; and further, the order picking is assigned in accordance with a picker's preference.
INTRODUCTION
SYNOPSIS:
An Equity Trading Portfolio Manager system that permits a customer to submit online orders for items and/or services from a store that serves both walk-in customers and online customers. The system presents an online display of an order cutoff time and an associated delivery window for items selected by the customer. The system accepts the customer's submission of a purchase order for the item in response to a time of submission being before the order cutoff time. The Equity Trading Portfolio Managersystem does not settle with a credit supplier of the customer until the item selected by the customer is picked from inventory but before it is delivered. Therefore, the customer can go online and make changes to the order. In addition, available service windows are presented to the customer as a function of customer selected order and service types; and further, the order picking is assigned in accordance with a picker's preference.
PROJECT DESCRIPTION:
The fundamental nature of grocery shopping makes it an interesting domain for intelligent mobile assistants. Even though the central role of shopping lists is widely recognized, relatively little attention has been paid to facilitating shopping list creation and management. In this paper we introduce a predictive text input technique that is based on association rules and item frequencies. We also describe an interface design for integrating the predictive text input with a web-based mobile shopping assistant. In a user study we compared two interfaces, one with text input support and one without. Our results indicate that, even though shopping list entries are typically short, our technique makes text input significantly faster, decreases typing error rates and increases overall user satisfaction.
SYSTEM STUDY
Existing System:
Any Company’s ultimate aim is to improve their products quality plus growth in the market. To do this, the company should possess goodwill of their customers and also good communication with the workers. Maintaining records for the purchasing activity of customers, staff details, and staff daily sales report is too hard and manual and prone to errors. Even when we develop software that will be available only in a stand alone system so the management people can’t view their day to day revenue if they were in out station .These are the main disadvantages we face in the existing System. We propose a solution to this problem which leads to the development of “Customer Relationship Management” (CRM).
Proposed System:
Customer Relationship Management (CRM), having a lot of attracting features and it draws the attentions of many big companies because of its elegant features. This is brings an easy way of communication with customers and workers and also helps the customers to view the products online and purchase in a shop. The company would directly get the feedback right from the customers and can thus maintain an excellent status amidst customers. It also have a easy to watch workers sales report. Customers can visit the website for viewing product and their features, after which they can choose their product and can come to the shop for purchasing or can book their products online in case of emergency. This reduces the traveling conveyances and even save the valuable time of the customers. Through this proposed system company can gain profit and good relationship among customers which serves as the valuable assets of the companies.
ECONOMICAL FEASIBILITY:
This study is carried out to check the economic impact that the system will have on the organization. The amount of fund that the company can pour into the research and development of the system is limited. The expenditures must be justified. Thus the developed system as well within the budget and this was achieved because most of the technologies used are freely available. Only the customized products had to be purchased.
TECHNICAL FEASIBILITY:
This study is carried out to check the technical feasibility, that is, the technical requirements of the system. Any system developed must not have a high demand on the available technical resources. This will lead to high demands on the available technical resources. This will lead to high demands being placed on the client. The developed system must have a modest requirement, as only minimal or null changes are required for implementing this system.
SOCIAL FEASIBILITY:
The aspect of study is to check the level of acceptance of the system by the user. This includes the process of training the user to use the system efficiently. The user must not feel threatened by the system, instead must accept it as a necessity. The level of acceptance by the users solely depends on the methods that are employed to educate the user about the system and to make him familiar with it. His level of confidence must be raised so that he is also able to make some constructive criticism, which is welcomed, as he is the final user of the system
SYSTEM DESIGN
Design is multi-step process that focuses on data structure software architecture, procedural details, (algorithms etc.) and interface between modules. The design process also translates the requirements into the presentation of software that can be accessed for quality before coding begins.
Computer software design changes continuously as new methods; better analysis and broader understanding evolved. Software Design is at relatively early stage in its revolution.
Therefore, Software Design methodology lacks the depth, flexibility and quantitative nature that are normally associated with more classical engineering disciplines. However techniques for software designs do exist, criteria for design qualities are available and design notation can be applied.
WHITE BOX TESTING:
The code testing strategy checks for the correctness of every statement in the program. To follow this testing strategy, there should be test cases that result in execution of every instruction in the program or module; that is every path in the program is tested. The test cases should guarantee that independent paths within a module are executed at least once.
BLACK BOX TESTING:
To perform black box testing, the analyst examines the specifications stating what the program or module should do and how it should perform under various conditions. Then, test cases are developed for various conditions and submitted for processing. By examining the results, the analyst can examine whether the program performs according the specified requirements.
Integration Testing:
The strategies for integrating software component product include the bottom-up strategy, the top-down strategy and send which strategy careful planning and scheduling are required. To ensure that modules will be available for integration into the evolving software product when needed. The integration strategy dictates the order in which modules must be available & thus exerts a strong influence on the order in which modules are written, debugged and unit tested.
Validation Testing:
At the culmination of integration testing, software is completely assembled as a package. Interfacing errors have been uncovered and corrected and a final series of software test-validation testing begins. Validation testing can be defined in many ways, but a simple definition is that validation succeeds when the software functions in manner that is reasonably expected by the customer. Software validation is achieved through a series of black box tests that demonstrate conformity with requirement. After validation test has been conducted, one of two conditions exists.
SOFTWARE DESCRIPTION
Introduction to .net framework
NET (dot-net) is the name Microsoft gives to its general vision of the future of computing, the view being of a world in which many applications run in a distributed manner across the Internet. We can identify a number of different motivations driving this vision.
Firstly, distributed computing is rather like object oriented programming, in that it encourages specialized code to be collected in one place, rather than copied redundantly in lots of places. There are thus potential efficiency gains to be made in moving to the distributed model.
Secondly, by collecting specialized code in one place and opening up a generally accessible interface to it, different types of machines (phones, handhelds, desktops, etc.) can all be supported with the same code. Hence Microsoft's 'run-anywhere' aspiration.
ADO .NET
Most applications need data access at one point of time making it a crucial component when working with applications. Data access is making the application interact with a database, where all the data is stored. Different applications have different requirements for database access. VB .NET uses ADO .NET (Active X Data Object) as it's data access and manipulation protocol which also enables us to work with data on the Internet. Let's take a look why ADO .NET came into picture replacing ADO.
Evolution of ADO.NET
The first data access model, DAO (data access model) was created for local databases with the built-in Jet engine which had performance and functionality issues. Next came RDO (Remote Data Object) and ADO (Active Data Object) which were designed for Client Server architectures but soon ADO took over RDO. ADO was a good architecture but as the language changes so is the technology. With ADO, all the data is contained in a record set object which had problems when implemented on the network and penetrating firewalls.
Why ADO.NET?
To cope up with some of the problems mentioned above, ADO .NET came into existence. ADO .NET addresses the above mentioned problems by maintaining a disconnected database access model which means, when an application interacts with the database, the connection is opened to serve the request of the application and is closed as soon as the request is completed. Likewise, if a database is Updated, the connection is opened long enough to complete the Update operation and is closed. By keeping connections open for only a minimum period of time, ADO .NET conserves system resources and provides maximum security for databases and also has less impact on system performance. Also, ADO .NET when interacting with the database uses XML and converts all the data into XML format for database related operations making them more efficient.
Buffer management
SQL Server buffers pages in RAM to minimize disc I/O. Any 8 KB page can be buffered in-memory, and the set of all pages currently buffered is called the buffer cache. The amount of memory available to SQL Server decides how many pages will be cached in memory. The buffer cache is managed by the Buffer Manager. Either reading from or writing to any page copies it to the buffer cache. Subsequent reads or writes are redirected to the in-memory copy, rather than the on-disc version. The page is updated on the disc by the Buffer Manager only if the in-memory cache has not been referenced for some time. While writing pages back to disc, asynchronous I/O is used whereby the I/O operation is done in a background thread so that other operations do not have to wait for the I/O operation to complete. Each page is written along with its checksum when it is written. When reading the page back, its checksum is computed again and matched with the stored version to ensure the page has not been damaged or tampered with in the mean time.
Logging and Transaction
SQL Server ensures that any change to the data is ACID-compliant, i.e., it uses transactions to ensure that any operation either totally completes or is undone if fails, but never leave the database in an intermediate state. Using transactions, a sequence of actions can be grouped together, with the guarantee that either all actions will succeed or none will. SQL Server implements transactions using a write-ahead log. Any changes made to any page will update the in-memory cache of the page, simultaneously all the operations performed will be written to a log, along with the transaction ID which the operation was a part of. Each log entry is identified by an increasing Log Sequence Number (LSN) which ensure that no event overwrites another. SQL Server ensures that the log will be written onto the disc before the actual page is written back. This enables SQL Server to ensure integrity of the data, even if the system fails. If both the log and the page were written before the failure, the entire data is on persistent storage and integrity is ensured. If only the log was written (the page was either not written or not written completely), then the actions can be read from the log and repeated to restore integrity. If the log wasn't written, then also the integrity is maintained, even though the database is in a state when the transaction as if never occurred. If it was only partially written, then the actions associated with the unfinished transaction are discarded. Since the log was only partially written, the page is guaranteed to have not been written, again ensuring data integrity. Removing the unfinished log entries effectively undoes the transaction. SQL Server ensures consistency between the log and the data every time an instance is restarted.
Concurrency and locking
SQL Server allows multiple clients to use the same database concurrently. As such, it needs to control concurrent access to shared data, to ensure data integrity - when multiple clients update the same data, or clients attempt to read data that is in the process of being changed by another client. SQL Server provides two modes of concurrency control: pessimistic concurrency and optimistic concurrency. When pessimistic concurrency control is being used, SQL Server controls concurrent access by using locks. Locks can be either shared or exclusive. Exclusive lock grants the user exclusive access to the data - no other user can access the data as long as the lock is held. Shared locks are used when some data is being read - multiple users can read from data locked with a shared lock, but not acquire an exclusive lock. The latter would have to wait for all shared locks to be released. Locks can be applied on different levels of granularity - on entire tables, pages, or even on a per-row basis on tables. For indexes, it can either be on the entire index or on index leaves.
Data retrieval
The main mode of retrieving data from an SQL Server database is querying for it. The query is expressed using a variant of SQL called T-SQL, a dialect Microsoft SQL Server shares with Sybase SQL Server due to its legacy. The query declaratively specifies what is to be retrieved. It is processed by the query processor, which figures out the sequence of steps that will be necessary to retrieve the requested data. The sequence of actions necessary to execute a query is called a query plan. There might be multiple ways to process the same query. For example, for a query that contains a join statement and a select statement, executing join on both the tables and then executing select on the results would give the same result as selecting from each table and then executing the join, but result in different execution plans. In such case, SQL Server chooses the plan that is supposed to yield the results in the shortest possible time. This is called query optimization and is performed by the query processor itself.
CONCLUSION
Although there are numerous advantages of shopping online, to the society as well as individuals, such as 24-hr access and absence of vendors, some negative concerns of Equity Trading Portfolio Managercan be reduced by efforts of Web makers. In addition to retail shops, thoughtful improvements of Web sites with consideration on consumers' persepective, Equity Trading Portfolio Managerwill gain respect and trust by consumers and will be much more widely used than today to benefit both the society and individuals. Eventully, almost all consumers will adapt to the Equity Trading Portfolio Managertechnology and will be able to enjoy the convenient, innovative way of shopping.