Distributed Database Systems (DDBS)

Choose a study mode

Play Quiz
Study Flashcards
Spaced Repetition
Chat to Lesson

Podcast

Play an AI-generated podcast conversation about this lesson

Questions and Answers

Relate both columns according to the description and its definition: Processing, function, data and control are:

Processing, function, data and control are: = Part of what is distributed Its goal is to integrate operational data this mean not centralization. For the above, computer network technology is oriented to connect distributed operational tasks. = DDBS motivation This is the result of the union of two technologies: << Database Systems >> and << Computer Network >>> = A Distributed Database System (DDBS) Web apps, e-commerce, manufacturing and cloud computing are computer technologies of: = Distributed architectures A number of autonomous processing elements (not necessarily homogeneous) that are interconnected by a computer network and that cooperate in performing their assigned tasks. = Distributed Computing

Relate both columns according to the description and its definition: Is a "collection of files" individually stored at each node of a computer network.

Is a "collection of files" individually stored at each node of a computer network. = Distributed file system This resides at one of the nodes of a network of computers - this is a centralized database on a network node. = A database system This is a collection of multiple, logically interrelated databases distributed over a computer network. = A distributed database (DDB) Processors at same sites are interconnected by a computer network. = A multiprocessor system This is the software that manages the DDB and provides an access mechanism that makes this distribution transparent to the users. = A distributed database management system

Relate both columns according to the description and its definition: The main characteristic of this delivery is that the arrival of new data items or updates to existing data items are carried out at a server without notification to clients.

The main characteristic of this delivery is that the arrival of new data items or updates to existing data items are carried out at a server without notification to clients. = Pull-only delivery Delivery modes, delivery frequency and communication methods are: = Are ortogonal dimensions into data delivery alternatives. The combination of data delivery alternatives provide a: = Rich design space for distributed systems This is related to how data is delivery from sites where they are stored to where the query is posed. = Data delivery The main difficulty of this approach is in deciding which data would be of common interest, and when to send them to clients. = Push-only delivery

Relate both columns according to the description and its definition: In this delivery, data are sent from servers whenever certain conditions installed by clients in their profiles are satisfied.

<p>In this delivery, data are sent from servers whenever certain conditions installed by clients in their profiles are satisfied. = Conditional delivery In this delivery is performed mostly in a pure pull-based system. = Irregular delivery In this delivery, the intervals can be defined by system default or by clients using their profiles. = Periodic delivery In this method, as the name implies, the server sends data to a number of clients. = One-to-many communication In this method, the communication is performed from a server to a client in a one-to-one fashion; the server sends data to one client using a particular delivery mode with some frequency. = Unicast communication In this mode, the continuous or continual query approach presents one possible way of combining the pull and push modes. = Hybrid delivery</p> Signup and view all the answers

Relate both columns according to the description and its definition: The first one refers to the immunity of user application to changes in the logical structure (scheme) of the database. The second one, deals with hiding the details of the storage structure from user applications.

<p>The first one refers to the immunity of user application to changes in the logical structure (scheme) of the database. The second one, deals with hiding the details of the storage structure from user applications. = Logical and physical data transparency This is a fundamental form of transparency, this refers to the immunity of user's applications to changes in the definition and organization of data, and vice versa. = Data independence The advantage of this system is the high level of support that it provides for the development of complex applications. = Fully transparent DBMS This is concerned with the user protection from the operational details of the network. = Network transparency This refers to separation of the higher-level semantics of a system from lower-level implementation issues. = Transparency promise</p> Signup and view all the answers

Relate both columns according to the description and its definition: This refers to the fact that the command used to perform a task is independent of both the location of the data and the system on which an operation is carried out.

<p>This refers to the fact that the command used to perform a task is independent of both the location of the data and the system on which an operation is carried out. = Location transparency For performance, reliability, and availability reasons, it is usually desirable to be able to distribute data in this transparency fashion across machines on a network. = Replication transparency Here, each sub-relation is defined as a subset of attributes (columns) of the original relation. = Vertical fragmentation This means that a unique name is provided for each object in the database. = Naming transparency In this transparency it is commonly desirable to divide each database relation into smaller fragments and treat each fragment as a separate database object. = Fragmentation transparency Here, a relation is partitioned into a set of sub-relations each of which have a subset of tuples (rows) of the original relation. = Horizontal fragmentation</p> Signup and view all the answers

Relate both columns according to the description and its definition: They are protocols that are capable of perform commits operations and recovery unfinished transactions.

<p>They are protocols that are capable of perform commits operations and recovery unfinished transactions. = Commit protocols It consists in the transformation of a consistent database state to another consistent database state even when several such transactions are executed concurrently. = Concurrency transparency Distributed DBMS are intended to improve reliability since they have replicated components and, thereby eliminates: = The single point of failure. They are protocols that control simultaneous transactions in distributed databases. = Distributed concurrency control protocols It happens when there are atomic failures during several transactions carried out at same time. = Failure atomicity This is a basic unit of consistent and reliable computing, consisting of a sequence of database operations executed as an atomic action. = A transaction</p> Signup and view all the answers

Relate both columns according to the description and its definition: They are mechanism to improve reliability in big size distributed databases (very common in NoSQL databases)

<p>They are mechanism to improve reliability in big size distributed databases (very common in NoSQL databases) = Data replication This results from the ability to execute multiple queries at the same time. = Inter-query parallelism A DBMS fragments the conceptual database, enabling data to be stored close to its points of use (also called data localization). = Proximity of data to its points of use Protocols designed to perform replication tasks considering principles of distributed computing. = Replication protocols This is achieved by breaking up a single query into several subqueries each of which is executed at a different site, accessing a different part of the distributed database. = Intra-query parallelism This is the capacity of allow a database scaling. = System expansion</p> Signup and view all the answers

Relate both columns according to the description and its definition: A system with proper support for database operations and dichotomy between general purpose processing requirements and database processing requirements are part of:

<p>A system with proper support for database operations and dichotomy between general purpose processing requirements and database processing requirements are part of: = Operating System Support How to distribute the database, whether is replicated &amp; non-replicated database distribution and the related problem in directory management are aspects of: = Distributed Database Design Synchronization of concurrent accesses, consistency and isolation of transactions' effects and deadlock management are aspects of: = Concurrency control How to make the system resilient to failures, atomicity and durability are aspects of: = System reliability This defines the structure of the system components identified, their functions, their interrelationships and interactions between components defined. = The structure of the system architecture Distributed Multidatabase Systems, more probable scenario and parallel issues are part of: = Open Systems and Interoperability</p> Signup and view all the answers

Relate both columns according to the description and its definition: The ability of a component DBMS to decide on issues related to its own design is:

<p>The ability of a component DBMS to decide on issues related to its own design is: = Design autonomy The ability of a component DBMS to execute local operations in any manner it wants to. = Execution autonomy Whether the components of the system are located on the same machine or not is part of: = Distribution dimension Horizontal and vertical scaling of resources, better price/performance on client machines and the capacity of client access to remote data (via standards) are: = Advantages of Client-Server architectures Hardware, communications, operating system data model, query language and transaction management algorithms are part of: = Heterogeneity distribution The ability of a component DBMS to decide whether and how to communicate with other DBMS. = Communication autonomy In this architecture we found out several database servers integrated to application servers through a network communication. = Distributed database server architecture.</p> Signup and view all the answers

Flashcards

Distributed Computing

Autonomous processing elements interconnected by a computer network that cooperate to perform tasks.

Distributed Database (DDB)

A collection of multiple, logically interrelated databases distributed over a computer network.

Distributed Database Management System

Software that manages the DDB and provides an access mechanism that makes this distribution transparent to users.

Conditional delivery

Data sent from servers when specific conditions set by clients are met.

Signup and view all the flashcards

Periodic delivery

Data is sent from servers with intervals defined by system default or client profiles.

Signup and view all the flashcards

Transparency promise

Separation of higher-level semantics of a system from lower-level implementation.

Signup and view all the flashcards

Horizontal fragmentation

A relation is partitioned into a set of sub-relations each of which have a subset of tuples (rows) of the original relation.

Signup and view all the flashcards

Distributed concurrency control protocols

Protocols that control simultaneous transactions in distributed databases.

Signup and view all the flashcards

System expansion

Scaling a database by distributing the load across multiple machines.

Signup and view all the flashcards

Distributed Database Design

How to distribute the database, whether it is replicated & non-replicated database distribution and the related problem in directory management.

Signup and view all the flashcards

Study Notes

  • Processing, function, data, and control are parts of what is distributed.
  • The goal of integrating operational data without centralization is a motivation for DDBS.
  • The union of Database Systems and Computer Networks results in a Distributed Database System (DDBS).
  • Web apps, e-commerce, manufacturing, and cloud computing are computer technologies of distributed architectures.
  • Distributed Computing involves autonomous processing elements interconnected by a computer network cooperating to perform tasks.
  • A distributed file system is a collection of files individually stored at each node of a computer network.
  • A database system resides at one of the nodes of a network of computers and is a centralized database on a network node.
  • A distributed database (DDB) is a collection of multiple, logically interrelated databases distributed over a computer network.
  • Processors at the same sites are interconnected by a multiprocessor system.
  • A distributed database management system manages the DDB and provides transparent access for users.
  • Pull-only delivery is characterized by the arrival of new or updated data at a server without client notification.
  • Delivery modes, delivery frequency, and communication methods are orthogonal dimensions in data delivery alternatives.
  • The combination of data delivery alternatives provides a rich design space for distributed systems.
  • Data delivery relates to how data is delivered from storage sites to query locations.
  • The main difficulty of push-only delivery is deciding which data would be of common interest and when to send it to clients.
  • In conditional delivery, data is sent from servers when certain conditions installed by clients in their profiles are satisfied.
  • Irregular delivery is performed mostly in a pure pull-based system.
  • In periodic delivery, intervals can be defined by system default or by clients using their profiles.
  • One-to-many communication involves the server sending data to a number of clients.
  • Unicast communication is performed from a server to a client in a one-to-one fashion, with the server sending data using a specific delivery mode and frequency.
  • Hybrid delivery, i.e. continuous or continual query, combines pull and push modes.
  • Logical and physical data transparency refers to the immunity of user applications to changes in the logical structure (scheme) of the database, and hiding storage structure details.
  • Data independence refers to the immunity of user's applications to changes in the definition and organization of data, and vice versa.
  • A fully transparent DBMS provides a high level of support for the development of complex applications.
  • Network transparency protects the user from the operational details of the network.
  • Transparency promise refers to separating the higher-level semantics of a system from lower-level implementation issues.
  • Location transparency means the command used to perform a task is independent of the location of the data and the system on which an operation is carried out.
  • Replication transparency allows to distribute data for performance, reliability, and availability across machines on a network.
  • Vertical fragmentation involves each sub-relation being defined as a subset of attributes (columns) of the original relation.
  • Naming transparency means that a unique name is provided for each object in the database.
  • Fragmentation transparency divides each database relation into smaller fragments and treats each fragment as a separate database object.
  • Horizontal fragmentation involves a relation partitioned into a set of sub-relations, each with a subset of tuples (rows) of the original relation.
  • Commit protocols are protocols that are capable of performing commit operations and recovering unfinished transactions.
  • Concurrency transparency consists in the transformation of a consistent database state to another consistent database state even when several such transactions are executed concurrently.
  • Distributed DBMS are meant to improve reliability since they have replicated components, thereby eliminating the single point of failure.
  • Distributed concurrency control protocols are protocols that control simultaneous transactions in distributed databases.
  • Failure atomicity happens when there are atomic failures during several transactions carried out at the same time.
  • A transaction is a basic unit of consistent and reliable computing, consisting of a sequence of database operations executed as an atomic action.
  • Data replication is a mechanism to improve reliability in big size distributed databases (very common in NoSQL databases).
  • Inter-query parallelism results from the ability to execute multiple queries at the same time.
  • Proximity of data to its points of use occurs when a DBMS fragments the conceptual database, enabling data to be stored close to its points of use (also called data localization).
  • Replication protocols are designed to perform replication tasks considering principles of distributed computing.
  • Intra-query parallelism is achieved by breaking up a single query into several subqueries, each of which is executed at a different site, accessing a different part of the distributed database.
  • System expansion is the capacity to allow a database scaling.
  • A system with proper support for database operations and dichotomy between general purpose processing requirements and database processing requirements are part of Operating System Support.
  • How to distribute the database, whether it is replicated or non-replicated database distribution, and the related problem in directory management, are aspects of Distributed Database Design.
  • Synchronization of concurrent accesses, consistency and isolation of transactions' effects, and deadlock management are aspects of Concurrency control.
  • How to make the system resilient to failures, atomicity, and durability are aspects of System reliability.
  • The structure of the system architecture defines the structure of the system components identified, their functions, their interrelationships, and interactions between components defined.
  • Distributed Multidatabase Systems, more probable scenarios, and parallel issues are part of Open Systems and Interoperability.
  • The ability of a component DBMS to decide on issues related to its own design is Design autonomy.
  • The ability of a component DBMS to execute local operations in any manner it wants to is Execution autonomy.
  • Whether the components of the system are located on the same machine or not is part of Distribution dimension.
  • Horizontal and vertical scaling of resources, better price/performance on client machines, and the capacity of client access to remote data (via standards) are Advantages of Client-Server architectures.
  • Hardware, communications, operating system data model, query language, and transaction management algorithms are part of Heterogeneity distribution.
  • The ability of a component DBMS to decide whether and how to communicate with other DBMS is Communication autonomy.
  • In Distributed database server architecture is where several database servers integrated to application servers through a network communication.

Studying That Suits You

Use AI to generate personalized quizzes and flashcards to suit your learning preferences.

Quiz Team

Related Documents

More Like This

Use Quizgecko on...
Browser
Browser