Click here to close now.

Welcome!

Java Authors: Liz McMillan, Trevor Parsons, Pat Romanski, Plutora Blog, Harry Trott

Related Topics: Java

Java: Article

The Challenges of Developing Distributed Java Applications

The Challenges of Developing Distributed Java Applications

In just a few years the Java language and platform has become the technical approach of choice for building complex, distributed and Web-enabled applications across the enterprise. Thanks to its cross-platform runtime environment, object-oriented development model, and facilities for working with object request brokers and other code components, Java is well equipped for building such applications.

Java enables software developers to provide seamless communication and application access to the rapidly growing world of Internet computers and communications devices, from UNIX servers and PCs to cell phones and beyond. IDC Research estimates that the Java products market will grow by 85% annually through 2004.

Java applications - especially those designed to work with other applications and components (often written in other languages) in a distributed environment - have different development requirements than traditional applications. For example, Java's execution model virtually eliminates traditional memory errors, but can introduce performance problems stemming from poor resource utilization. For those used to addressing traditional programming errors and other issues, Java's unique execution model and language characteristics may make building error-free and efficient applications more difficult.

The distributed nature of many Java applications can also make it difficult to pinpoint performance issues or diagnose programming errors. As a result, Java developers need software development tools and techniques for viewing and analyzing code execution on multiple systems on the network and from multiple code bases.

Issues Surrounding the Development of Distributed Java
Java is unique - it's a mainstream programming language that works like no other. Its rules aren't well understood yet by many application developers. Part of the reason for this is that its capabilities and limitations haven't been fully explored.

As a result, while many of the problems may be similar, recognizing them and knowing what to do when you find them remains challenging, even to experienced Java developers. What follows are just a few of the development issues and what they mean for Java.

Performance
Performance is a concern of applications written in any language. Most programmers are familiar with common performance issues using a conventional language such as C with a stand-alone or even a client/server application. Such issues often involve improper allocation, deallocation of memory and poor use of system APIs.

These aren't even characteristics of the Java language. For example, relating Java code to how the JVM manages memory is difficult and error-prone. However, it's vital to improve Java performance because its execution model has additional overhead that tends to degrade performance more than native applications do.

In addition to typical performance requirements and issues surrounding traditional stand-alone applications, distributed Java applications must contend with problems surrounding the interactions between components running on different systems. Performance problems may manifest themselves in unexpected ways or appear to be caused by different parts of the code other than the actual problem area. Identifying and locating performance bottlenecks rapidly is a significant challenge in distributed application development.

Reliability
Enterprise Java applications, especially distributed ones, are often mission-critical in nature: all aspects of the application must work perfectly at all times. Developers and development managers must be able to gauge the reliability of their applications accurately. While the characteristics of the language tend to make Java applications less error-prone, there are still plenty of ways to introduce runtime errors.

With distributed Java applications the reliability equation is even more difficult. It means assessing not only the individual applications but also the components as they interact. Java also makes it possible to write highly threaded applications that make sense in a distributed environment. But using threads means that problems with resource contention and deadlock are much greater.

Testing
Java applications face the same testing problems as traditional ones. They must be tested as thoroughly as possible before fielded, and developers should have a good idea of the extent of the test coverage before certifying an application. Distributed software systems written in Java, however, are extraordinarily difficult to test and debug. Because components reside on different computers and must work together perfectly for the application to work properly, all components must be tested simultaneously.

Memory Management
Since JVMs perform all the memory management tasks for applications, understanding the memory usage of the underlying platform and influencing memory allocation and use to affect performance is difficult to do and not intuitive. That's especially true because developers have no easy way of determining the relationship between code and the underlying memory use. Analyzing underlying memory use is a key component of building efficient applications (see Figure 1).

Software Tools Can Make Java Transparent
Several of the integrated development environments for Java are quite good, combining visual development, context-sensitive editing, JIT compilation and runtime debugging. What they lack, however, is the ability to determine the efficiency and reliability of the application, especially if it's distributed across several servers.

Most developers do without such tools, due in part to the relatively poor selection. There are many different Java-oriented development environments, but few tools to move code beyond the development stage. A few tools, such as the Compuware DevPartner for Java suite, combine components that evaluate performance issues, examine memory usage, analyze threads and track testing progress (see Figure 2).

Many Java developers fail to recognize that it's not enough for Java applications to be debugged within the development environment because of several myths regarding the use of Java as a development language and environment. One myth is that the VM eliminates programming errors and bugs. While direct memory errors aren't usually possible in Java development, it's still possible for Java applications to contain serious errors that affect the proper operation of the program.

Another myth is that the application developer has no control over the performance of the software since the VM manages the low-level details that determine how fast the code runs. How developers use specific language instructions can have a significant impact on application performance. Often a few simple changes can greatly improve performance if developers know their underlying effect in advance.

Java development tools assist and accelerate the development of reliable, high-performance applications, especially distributed ones. They go beyond the traditional development environments to include components that help make applications more reliable and efficient. When choosing a set of tools to supplement those found in development environments and improve the reliability and performance of Java applications, developers and development teams should take the following characteristics into account.

  • Support for multiple, unmodified VMs: Different operating systems often use different JVMs with different performance and behavior characteristics. In addition, some Java tools require the use of specially modified VMs that may not represent the characteristics of production systems.

    To obtain accurate information to improve the performance and reliability of distributed Java and mixed-language applications, developers should select tools that run in the actual deployment environment. This way they can be certain that the behavior observed and analyzed during development and testing will be the same once the application is deployed.

  • Minimal impact on Java runtime resources: During application testing and analysis, the Java development tools that are used can often be intrusive - their presence may influence the test results. If developers attempt to deploy the tools for further testing and tuning in the production environment, they may also encounter unacceptable performance degradation due to high resource utilization.

    If developers need accurate information on resource utilization, performance and system load, they should consider software tools that don't use large amounts of Java runtime resources. Tools with minimal Java and computing resource impact are more likely to provide the accurate information needed to improve the performance and reliability of distributed applications.

  • Support for Web technologies such as JavaScript and Active Server Pages (ASP): Large distributed applications that use Java often go beyond it to include many different software technologies, including JavaScript and ASP. Web software technologies can be a source of reliability and performance problems, and software tools that are unable to test for and identify these problems are of limited use.

    Application developers need tools that work with whatever software technologies they're using. Multilanguage tools, especially those that support industry-standard Web languages, provide the range of capabilities needed to find performance and reliability problems anywhere in the application (see Figure 3).

  • Ability to track Java memory utilization to the function and line level: Knowing you have a performance or reliability problem isn't any good unless you're able to pinpoint the source rapidly and easily. The more precise the diagnosis and analysis, the faster application developers can address the problem.

    Java application development tools should focus as specifically as possible on the exact location of a performance bottleneck or software error, especially for large, distributed applications. Developers working on deadlines need the most exacting information possible from their tools.

  • Ability to find thrashing and starvation conditions graphically: Some of the toughest software problems to find are those that involve resource starvation and code thrashing. These error conditions don't prevent the application from running; however, they cause severe performance bottlenecks and may even cause the application to hang while running.

    The problems are especially prevalent in Java applications in which multiple running threads contend for limited virtual machine resources. To ensure the reliability of deployed applications, Java developers require software tools that enable them to identify and locate the complex combination of conditions that can cause resource starvation and thrashing.

  • Ability to measure code base stability: Rapidly changing code during debugging and testing usually means that the application may be unreliable or needs additional testing before deployment. Conversely, an application with few changes to the code base during debugging and testing will more likely be fully tested and stable.

    It's important, therefore, for application developers to understand how much and how rapidly their application code is changing during the latter stages of the development process. An application development tool should measure the stability of the code base to enable both the development team and the managers to determine when the application can be used reliably in production.

  • Support for multiple operating systems: Mixed operating environments are the rule rather than the exception. Even if developers code on one platform, it's increasingly likely that the application will be deployed across several different types of systems. This is especially true of Java-based applications, which are designed to run unmodified on multiple operating systems.

    Application development tools have to support multiple development and runtime environments to eliminate the expense of purchasing different tool sets for different platforms, and to reduce the need for developer training on multiple tool sets and platforms.

Delivering Higher Quality Java Applications
Today's Web-enabled, distributed applications combine many different technologies and are prone to performance and reliability problems. Software developers using Java technology can spend a substantial amount of time trying to resolve these problems, leading to schedule delays and applications with ongoing problems.

Most of the Java development environments available are excellent for writing small to medium-sized stand-alone applications. It's growing increasingly difficult to write applications that are large or distributed, and to work with legacy components or databases. Performance and reliability issues overshadow the advantages of rapid, object-oriented development.

Visualizing these types of problems is an important aspect of debugging, tuning and testing applications because developers can quickly identify and localize the code responsible. If the developer can see where the deadlocked thread is, it's easier to pinpoint the resource that's deadlocked and its cause.

The kinds of problems that don't arise or are trivial in smaller applications take on critical importance in distributed processing. By using performance analyzers, memory profilers and thread inspectors such as those found in DevPartner Java Edition, distributed application developers can be sure their applications will run as expected.

Large-scale Java and distributed applications are easy to write but difficult to write well. To bridge the gap between mediocre or poorly performing applications and highly efficient ones, developers need to leverage software tools that help build reliable high-performance applications and components with Java technology. Such tools would allow Java developers to quickly and easily identify problems in key areas like runtime performance, memory utilization and multithreading.

More Stories By Peter Varhol

Peter Varhol is a Product Manager at Compuware Corporation, with graduate degrees in Mathematics and Computer Science, and has worked in software development and systems management. As a former college professor, he has substantial experience speaking on technical topics. He has written extensively on software development topics in a variety of trade magazines.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
The Internet of Things (IoT) is causing data centers to become radically decentralized and atomized within a new paradigm known as “fog computing.” To support IoT applications, such as connected cars and smart grids, data centers' core functions will be decentralized out to the network's edges and endpoints (aka “fogs”). As this trend takes hold, Big Data analytics platforms will focus on high-volume log analysis (aka “logs”) and rely heavily on cognitive-computing algorithms (aka “cogs”) to make sense of it all.
SYS-CON Events announced today that Vitria Technology, Inc. will exhibit at SYS-CON’s @ThingsExpo, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Vitria will showcase the company’s new IoT Analytics Platform through live demonstrations at booth #330. Vitria’s IoT Analytics Platform, fully integrated and powered by an operational intelligence engine, enables customers to rapidly build and operationalize advanced analytics to deliver timely business outcomes for use cases across the industrial, enterprise, and consumer segments.
The Internet of Everything (IoE) brings together people, process, data and things to make networked connections more relevant and valuable than ever before – transforming information into knowledge and knowledge into wisdom. IoE creates new capabilities, richer experiences, and unprecedented opportunities to improve business and government operations, decision making and mission support capabilities. In his session at @ThingsExpo, Gary Hall, Chief Technology Officer, Federal Defense at Cisco Systems, will break down the core capabilities of IoT in multiple settings and expand upon IoE for bo...
The explosion of connected devices / sensors is creating an ever-expanding set of new and valuable data. In parallel the emerging capability of Big Data technologies to store, access, analyze, and react to this data is producing changes in business models under the umbrella of the Internet of Things (IoT). In particular within the Insurance industry, IoT appears positioned to enable deep changes by altering relationships between insurers, distributors, and the insured. In his session at @ThingsExpo, Michael Sick, a Senior Manager and Big Data Architect within Ernst and Young's Financial Servi...
SYS-CON Events announced today that GENBAND, a leading developer of real time communications software solutions, has been named “Silver Sponsor” of SYS-CON's WebRTC Summit, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. The GENBAND team will be on hand to demonstrate their newest product, Kandy. Kandy is a communications Platform-as-a-Service (PaaS) that enables companies to seamlessly integrate more human communications into their Web and mobile applications - creating more engaging experiences for their customers and boosting collaboration and productiv...
From telemedicine to smart cars, digital homes and industrial monitoring, the explosive growth of IoT has created exciting new business opportunities for real time calls and messaging. In his session at @ThingsExpo, Ivelin Ivanov, CEO and Co-Founder of Telestax, shared some of the new revenue sources that IoT created for Restcomm – the open source telephony platform from Telestax. Ivelin Ivanov is a technology entrepreneur who founded Mobicents, an Open Source VoIP Platform, to help create, deploy, and manage applications integrating voice, video and data. He is the co-founder of TeleStax, a...
The industrial software market has treated data with the mentality of “collect everything now, worry about how to use it later.” We now find ourselves buried in data, with the pervasive connectivity of the (Industrial) Internet of Things only piling on more numbers. There’s too much data and not enough information. In his session at @ThingsExpo, Bob Gates, Global Marketing Director, GE’s Intelligent Platforms business, to discuss how realizing the power of IoT, software developers are now focused on understanding how industrial data can create intelligence for industrial operations. Imagine ...
The explosion of connected devices / sensors is creating an ever-expanding set of new and valuable data. In parallel the emerging capability of Big Data technologies to store, access, analyze, and react to this data is producing changes in business models under the umbrella of the Internet of Things (IoT). In particular within the Insurance industry, IoT appears positioned to enable deep changes by altering relationships between insurers, distributors, and the insured. In his session at @ThingsExpo, Michael Sick, a Senior Manager and Big Data Architect within Ernst and Young's Financial Servi...
Operational Hadoop and the Lambda Architecture for Streaming Data Apache Hadoop is emerging as a distributed platform for handling large and fast incoming streams of data. Predictive maintenance, supply chain optimization, and Internet-of-Things analysis are examples where Hadoop provides the scalable storage, processing, and analytics platform to gain meaningful insights from granular data that is typically only valuable from a large-scale, aggregate view. One architecture useful for capturing and analyzing streaming data is the Lambda Architecture, representing a model of how to analyze rea...
One of the biggest impacts of the Internet of Things is and will continue to be on data; specifically data volume, management and usage. Companies are scrambling to adapt to this new and unpredictable data reality with legacy infrastructure that cannot handle the speed and volume of data. In his session at @ThingsExpo, Don DeLoach, CEO and president of Infobright, will discuss how companies need to rethink their data infrastructure to participate in the IoT, including: Data storage: Understanding the kinds of data: structured, unstructured, big/small? Analytics: What kinds and how responsiv...
The 3rd International @ThingsExpo, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - is now accepting submissions to demo smart cars on the Expo Floor. Smart car sponsorship benefits include general brand exposure and increasing engagement with the developer ecosystem.
Since 2008 and for the first time in history, more than half of humans live in urban areas, urging cities to become “smart.” Today, cities can leverage the wide availability of smartphones combined with new technologies such as Beacons or NFC to connect their urban furniture and environment to create citizen-first services that improve transportation, way-finding and information delivery. In her session at @ThingsExpo, Laetitia Gazel-Anthoine, CEO of Connecthings, will focus on successful use cases.
Sensor-enabled things are becoming more commonplace, precursors to a larger and more complex framework that most consider the ultimate promise of the IoT: things connecting, interacting, sharing, storing, and over time perhaps learning and predicting based on habits, behaviors, location, preferences, purchases and more. In his session at @ThingsExpo, Tom Wesselman, Director of Communications Ecosystem Architecture at Plantronics, will examine the still nascent IoT as it is coalescing, including what it is today, what it might ultimately be, the role of wearable tech, and technology gaps stil...
Sensor-enabled things are becoming more commonplace, precursors to a larger and more complex framework that most consider the ultimate promise of the IoT: things connecting, interacting, sharing, storing, and over time perhaps learning and predicting based on habits, behaviors, location, preferences, purchases and more. In his session at @ThingsExpo, Tom Wesselman, Director of Communications Ecosystem Architecture at Plantronics, will examine the still nascent IoT as it is coalescing, including what it is today, what it might ultimately be, the role of wearable tech, and technology gaps stil...
When it comes to the Internet of Things, hooking up will get you only so far. If you want customers to commit, you need to go beyond simply connecting products. You need to use the devices themselves to transform how you engage with every customer and how you manage the entire product lifecycle. In his session at @ThingsExpo, Sean Lorenz, Technical Product Manager for Xively at LogMeIn, will show how “product relationship management” can help you leverage your connected devices and the data they generate about customer usage and product performance to deliver extremely compelling and reliabl...
SYS-CON Events announced today that SoftLayer, an IBM company, has been named “Gold Sponsor” of SYS-CON's 16th International Cloud Expo®, which will take place June 9-11, 2015 at the Javits Center in New York City, NY, and the 17th International Cloud Expo®, which will take place November 3–5, 2015 at the Santa Clara Convention Center in Santa Clara, CA. SoftLayer operates a global cloud infrastructure platform built for Internet scale. With a global footprint of data centers and network points of presence, SoftLayer provides infrastructure as a service to leading-edge customers ranging from ...
SYS-CON Events announced today that Open Data Centers (ODC), a carrier-neutral colocation provider, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place June 9-11, 2015, at the Javits Center in New York City, NY. Open Data Centers is a carrier-neutral data center operator in New Jersey and New York City offering alternative connectivity options for carriers, service providers and enterprise customers.
There’s Big Data, then there’s really Big Data from the Internet of Things. IoT is evolving to include many data possibilities like new types of event, log and network data. The volumes are enormous, generating tens of billions of logs per day, which raise data challenges. Early IoT deployments are relying heavily on both the cloud and managed service providers to navigate these challenges. Learn about IoT, Big Data and deployments processing massive data volumes from wearables, utilities and other machines.
The true value of the Internet of Things (IoT) lies not just in the data, but through the services that protect the data, perform the analysis and present findings in a usable way. With many IoT elements rooted in traditional IT components, Big Data and IoT isn’t just a play for enterprise. In fact, the IoT presents SMBs with the prospect of launching entirely new activities and exploring innovative areas. CompTIA research identifies several areas where IoT is expected to have the greatest impact.
Wearable devices have come of age. The primary applications of wearables so far have been "the Quantified Self" or the tracking of one's fitness and health status. We propose the evolution of wearables into social and emotional communication devices. Our BE(tm) sensor uses light to visualize the skin conductance response. Our sensors are very inexpensive and can be massively distributed to audiences or groups of any size, in order to gauge reactions to performances, video, or any kind of presentation. In her session at @ThingsExpo, Jocelyn Scheirer, CEO & Founder of Bionolux, will discuss ho...