1answer.
Ask question
Login Signup
Ask question
All categories
  • English
  • Mathematics
  • Social Studies
  • Business
  • History
  • Health
  • Geography
  • Biology
  • Physics
  • Chemistry
  • Computers and Technology
  • Arts
  • World Languages
  • Spanish
  • French
  • German
  • Advanced Placement (AP)
  • SAT
  • Medicine
  • Law
  • Engineering
vichka [17]
3 years ago
9

Select the correct answer.

Computers and Technology
1 answer:
serg [7]3 years ago
4 0

long travel between the sender and the receiver

You might be interested in
To obtain your class E learner license, youll need too _
k0ka [10]
<h3>                                           Answer:</h3>

  To obtain you  class E Learner License,  you'll need to <u>Pass a literacy Test</u>

<h2>         </h2><h2>                 Can you Brainliest me?</h2>
8 0
3 years ago
True or False: To create a function in python, you start with the keyword "def"
Butoxors [25]

Answer:

true i think

Explanation:

8 0
3 years ago
Read 2 more answers
Which dhcp client option is used to provide a list of domain names to use in name resolution queries?
Rama09 [41]
Name resolution is done using DNS servers. DHCP provides a client with a list of DNS servers via option 6, "DNS Server".
4 0
4 years ago
8. Explain what the following spredications of a monitor refer to their significance for the user and how
Maslowich

Answer:

A computer monitor is an output device that displays information in pictorial form. A monitor usually comprises the visual display, circuitry, casing, and power supply. The display device in modern monitors is typically a thin film transistor liquid crystal display (TFT-LCD) with LED backlighting having replaced cold-cathode fluorescent lamp (CCFL) backlighting. Previous monitors used a cathode ray tube (CRT). Monitors are connected to the computer via VGA, Digital Visual Interface (DVI), HDMI, DisplayPort, USB-C, low-voltage differential signaling (LVDS) or other proprietary connectors and signals.

Originally, computer monitors were used for data processing while television sets were used for entertainment. From the 1980s onwards, computers (and their monitors) have been used for both data processing and entertainment, while televisions have implemented some computer functionality. The common aspect ratio of televisions, and computer monitors, has changed from 4:3 to 16:10, to 16:9.

Modern computer monitors are easily interchangeable with conventional television sets and vice versa. However, as computer monitors do not necessarily include integrated speakers nor TV tuners (such as Digital television adapters), it may not be possible to use a computer monitor as a TV set without external components.[1]

History

Early electronic computers were fitted with a panel of light bulbs where the state of each particular bulb would indicate the on/off state of a particular register bit inside the computer. This allowed the engineers operating the computer to monitor the internal state of the machine, so this panel of lights came to be known as the 'monitor'. As early monitors were only capable of displaying a very limited amount of information and were very transient, they were rarely considered for program output. Instead, a line printer was the primary output device, while the monitor was limited to keeping track of the program's operation.[2]

Computer monitors were formerly known as visual display units (VDU), but this term had mostly fallen out of use by the 1990s.

Technologies

Further information: Comparison of CRT, LCD, Plasma, and OLED and History of display technology

Multiple technologies have been used for computer monitors. Until the 21st century most used cathode ray tubes but they have largely been superseded by LCD monitors.

Cathode ray tube

Main article: Cathode ray tube

The first computer monitors used cathode ray tubes (CRTs). Prior to the advent of home computers in the late 1970s, it was common for a video display terminal (VDT) using a CRT to be physically integrated with a keyboard and other components of the system in a single large chassis. The display was monochrome and far less sharp and detailed than on a modern flat-panel monitor, necessitating the use of relatively large text and severely limiting the amount of information that could be displayed at one time. High-resolution CRT displays were developed for the specialized military, industrial and scientific applications but they were far too costly for general use.

Some of the earliest home computers (such as the TRS-80 and Commodore PET) were limited to monochrome CRT displays, but color display capability was already a standard feature of the pioneering Apple II, introduced in 1977, and the speciality of the more graphically sophisticated Atari 800, introduced in 1979. Either computer could be connected to the antenna terminals of an ordinary color TV set or used with a purpose-made CRT color monitor for optimum resolution and color quality. Lagging several years behind, in 1981 IBM introduced the Color Graphics Adapter, which could display four colors with a resolution of 320 × 200 pixels, or it could produce 640 × 200 pixels with two colors. In 1984 IBM introduced the Enhanced Graphics Adapter which was capable of producing 16 colors and had a resolution of 640 × 350.[3]

By the end of the 1980s color CRT monitors that could clearly display 1024 × 768 pixels were widely available and increasingly affordable. During the following decade, maximum display resolutions gradually increased and prices continued to fall. CRT technology remained dominant in the PC monitor market into the new millennium partly because it was cheaper to produce and offered to view angles close to 180°.[4] CRTs still offer some image quality advantages[clarification needed] over LCDs but improvements to the latter have made them much less obvious. The dynamic range of early LCD panels was very poor, and although text and other motionless graphics were sharper than on a CRT, an LCD characteristic known as pixel lag caused moving graphics to appear noticeably smeared and blurry.

Explanation:

8 0
3 years ago
What are the advantages of block-based programming? What are the disadvantages?
krek1111 [17]

Answer:

Block-based languages tend to have a lot of commands to choose from, making memorizing commands unneeded, and therefore easy for beginners to start coding. However, it may be slower to code in a block-based language than one where a user can quickly input commands using their keyboard. With block based programming you don't have to worry about syntax errors which is helpful.

Explanation:

8 0
3 years ago
Read 2 more answers
Other questions:
  • A data structure used to bind an authenticated individual to a public key is the definition of ________.
    14·1 answer
  • In an open computer network such as the internet, hipaa requires the use of _____. in a closed system such as a local area netwo
    7·1 answer
  • Determine whether the relation represents a function. If it is a​ function, state the domain and range.
    6·1 answer
  • Jess and Ann have decided to exchange messages using the Caesar cipher. If the predetermined key is seven places to the left, th
    10·1 answer
  • You have a computer at home. The computer is connected to the Internet through a dial-up connection. Every time you connect to t
    14·1 answer
  • When the Hyper-V role is added to a Windows Server 2016 server, the hypervisor creates the individual environments, each of whic
    5·1 answer
  • Will give brainliest
    8·1 answer
  • What is the biggest problem with technology today?
    10·1 answer
  • What is fruit nursery?​
    8·2 answers
  • To create a cell reference in another spreadsheet you must need to start with _______ sign.
    14·1 answer
Add answer
Login
Not registered? Fast signup
Signup
Login Signup
Ask question!