Memory Computer Prototype

The computer has 160 T-bytes of memory, and HPE expects the architecture to allow memory to scale up to 4.096 yottabytes.

The memory is spread across 40 physical nodes that are interconnected using a high-performance fabric protocol.

The computer runs on an optimized Linux-based operating system running ThunderX2, Cavium’s flagship second-generation dual socket-capable ARMv8-A workload optimized System on a Chip.

It uses photonics/optical communication links, including the new HPE X1 photonics module.

The computer has software programming tools designed to take advantage of abundant persistent memory.

The technology is built for the big data era, HPE said.

One Size Fits All

“We think we’ve got an architecture that scales all the way from edge devices — the intelligent edge — through some bigger systems that are probably more in line with what we’ve built out as a prototype,” said HPE Fellow Andrew Wheeler, deputy director of Hewlett Packard Labs.

“If you think competitively against that scale of things, you’re going up against conventional supercomputers, which have their

How to get the magnets out of the hard drive

Hard Drive Magnets – Strong and Fun

Hard drive magnets can be heaps of fun. I take them out of my old hard drives and distribute them to my brothers. They can be useful for holding things together and are fun to fiddle with. They can be used to fix the book that was open on its spine and are great for competitions in seeing who can pull the magnets apart (this is extremely hard). The magnets are powerful enough to go through someone’s hand and maybe even two hands.

How to get the magnets out of the hard drive

The first step in getting some hard drive magnets is finding a hard drive that you no longer need or is broken in some way. Remember once you open it, the hard drive will no longer be usable. Firstly find all the screws and take them out. Some hard drives will have different screws meaning that you may have to drill through the screw or lever it using a flat screw driver.

Once you take all the screws out

Computer Software and Hardware

Definition of Computer Hardware and Software

When we talk about computer hardware, we mean the actual components of your computer. Such things as the computer’s motherboard, its CPU, the video card, the keyboard and mouse, these are all “hardware”.

The difference between computer software and hardware is that software refers to the coding and various programs that you have on your computer. These include your operating system (Windows etc), media players, Photoshop etc.


Computer hardware is usually multi-purpose in that it is able to perform lots of different tasks. For instance, your computer monitor doesn’t just display images on screen; it also shows videos, widgets and text. One difference between computer software and hardware is that software is normally only designed to perform one task.

Your media player for example, is only for accessing media like movies and songs. It cannot edit photos or browse the web. The only real exception to this is the operating system itself, which is a user-friendly interface designed to let you access all the other bits of software and files stored on your PC.

Scared of computers

I usually don’t categorize people into certain groups but in this circumstance I will. You are either one of two people. The first person sees the link ‘Computers are not scary’ and out of curiosity clicks the link. The second person is genuinely worried and maybe scared of computers and has clicked the link hoping to be reassured. Don’t worry! You are going to be reassured of the fact that computers aren’t scary.

Maybe you are a person that is very worried about computers and feels that if you click the wrong thing the whole computer will stop working or maybe even blow up. Maybe you find computers superstitious and feel that it will all of a sudden do something wrong. This page will hopefully reassure you of all your worries and make you feel a whole lot more comfortable in regards to them.

First of all, I would like to state that computers just don’t stop working. There is always a reason when a computer stops working. Maybe there is a computer virus or a program needs updating or some hardware had reached its use by date. You need to reassure yourself that if your

The notion of building computers

Transistors operate more like neurons began in the 1980s with Caltech professor Carver Mead. One of the core arguments behind what Mead came to call “neuromorphic” computing was that semiconductor devices can, when operated in a certain mode, follow the same physical rules as neurons do and that this analog behavior could be used to compute with a high level of energy efficiency.

The Mead group also invented a neural communication framework in which spikes are coded only by their network addresses and the time at which they occur. This was groundbreaking work because it was the first to make time an essential feature of artificial neural networks. Time is a key factor in the brain: It takes time for signals to propagate and membranes to respond to changing conditions, and time determines the shape of postsynaptic potentials.

Several research groups quite active today, such as those of Giacomo Indiveriat ETH Zurich and Kwabena Boahen at Stanford, have followed Mead’s approach and successfully implemented elements of biological cortical networks. The trick is to operate transistors below their turn-on threshold with extremely low currents, creating analog circuits that mimic neural behavior while consuming very little energy.

Further research in

The Coolest Thing Out of Computex

Computex this year, and that was sad for everything but my budget, because there was a ton of cool stuff announced at the show. Dell, HP and Lenovo showed off new designs that were both attractive and compelling. Mixed-reality headsets hit; based on Intel and Microsoft technology, they were far more affordable than the strong virtual reality stuff already in market (andsome aren’t bad looking). New core wars broke out, as AMD’s 16 Core Threadripper was challenged by Intel’s 18 core i9.

It seems that gaming was huge at Computex this year. The product — or the concept really — that stood out most to me was Nvidia’s Max-Q gaming laptop concept, which promises a gaming laptop with dimensions that would rival a MacBook Air.

I’ll focus on that this week and close with my product of the week: the new smartphone that Apple is working furiously to kill before it can be launched (which is why I immediately ordered one).

The Historic Gaming Laptop Problem

Here is the deal — I love to play games both at home and when I travel, and I tend to lock in on one

Computer graphics


Carol Luckhardt Redfield is a Professor of Computer Science and Graduate Program Director for Computer Science at St. Mary’s University in San Antonio, Texas. She was in the computer industry for 15 years before teaching at St. Mary’s. She is specialized in educational computer gaming. She completed her PhD from the University of Michigan in Computer Science and Engineering with her work on Artificial Intelligence and Gaming. She currently serves in committees with the San Antonio Space Society and Friends Meeting of San Antonio (Quakers).

The computer graphics class at St. Mary’s University focuses on the applications of computer graphics while learning graphic terms, theory of how graphic tools work, and common graphic creation tools similar to Microsoft Paint, Adobe Photoshop, Adobe Flash, and Adobe Dreamweaver. In the past, computer graphics classes were about how to make computer graphic tools and now the need has shifted to making graphics. Students in the class have to create a brand for themselves, a group, a company or an organization that they select. Inside of that brand, students have to create logos, brochure, business card, business stationary, animation file, and a website. The website incorporates all the graphic work that they did during the

The Linux Desktop

“Desktop environment” is the technical term for a typical, full-featured desktop — that is, the complete graphical layout of your system. Besides displaying your programs, the desktop environment includes accoutrements such as app launchers, menu panels and widgets.

In Microsoft Windows, the desktop environment consists of, among other things, the Start menu, the taskbar of open applications and notification center, all the Windows programs that come bundled with the OS, and the frames enclosing open applications (with a dash, square and X in the upper right corner).

There are many similarities in Linux.

The Linux Gnome desktop environment, for instance, has a slightly different design, but it shares all of the Microsoft Windows basics — from an app menu to a panel showing open applications, to a notification bar, to the windows framing programs.

Window program frames rely on a component for drawing them and letting you move and resize them: It’s called the “window manager.” So, as they all have windows, every desktop environment includes a window manager.

However, not every window manager is part of a desktop environment. You can run window managers by themselves, and there are reasons to consider doing just

The computer technology

The aim of this paper is to present a system for controlling the cursor position in a computer by human eye movements. Many modules have been developed to help the physical world interact with the digital world. Here we present a novel approach for Human Computer Interaction (HCI) where, we control cursor movement using a real time camera and color pointers. Our method is to use a camera and computer vision technology, such as image segmentation, background subtraction and color tracking, to control mouse tasks and we show how it can perform everything as current mouse devices can. A color pointer has been used for the object recognition and tracking, so as to implement the module without any physical contact with the system. The application has been created on matlab environment with operating system as windows 8.1. This method mainly focuses on the use of a web camera to develop a virtual human computer interaction device in a cost effective manner presents hands free interface between computer and human especially for physically disabled persons.

Human Computer Interaction (HCI); Eye ball movement; MATLAB; Web camera; Computer

As the computer technology is grow up, the importance of human computer interaction

The sad, pathetic personality of a computer hacker

Why don’t I just call this professional spammer and give him a piece of my mind? What’s his number? Who’s his supervisor? Should we get the cops involved?”

“The police can’t help Dad,” I replied. “They can’t catch a hacker.”

How do you explain to a senior citizen just entering the Digital Age that his life could be turned upside down in seconds by unseen, nameless forces that wreak havoc on computer novices? You know who you are. Some of you are so proud of your useless skills that you post YouTube tutorials detailing how to create a virus. The videos contain your voice but not your face. But even though you hide behind a cloak of secrecy, you are not entirely anonymous. I know things about you. In fact, I know your movements from the moment you wake up. Does any of this sound familiar?

You roll off your floor mattress whenever you feel like it. You have no alarm clock because you are unemployed and have no desire to change your work status. Having a job requires both motivation and people skills. You have neither.

With the touch

Computers Work

Have you ever just looked at a computer and go “how does that work”?  If you think about it, computers are quite amazing.  You press a few buttons and you can talk to your friend over the other side of the world, you can learn anything you want on the internet, you can listen to music, watch TV, write stories, make videos and do much much more.

The more you think about it the more amazing it is and the more crazy it is that a bunch of computer bits can do this.  So how do computers work?

Well I will try and answer this the best that I can.  Basically a computer can consist of two broad categories, hardware and software and the way a computer works is between these two working together. I will oultine them both briefly below.

Computer Hardware

If you have been reading any of this website, then you probably already understand what computer hardware is, because that is what this website is about. Like seriously,

So briefly: computer hardware is the physical computer that you actually see, this includes: the computer, monitor, keyboard, mouse, printer, speakers etc.  Inside the computer is more hardware,

Computer Hardware

 if you only have a laptop computer, you will still have them, only they are integrated into your computer. For more information about learning hardware click here, or otherwise just continue reading for a basic run down…

System Unit

This is the actual computer – All the other bits are known as peripheral devices. When discussing what is computer hardware, this is the most important example we can give.

The system unit, in the case of a desk top computer, is the big chunky thing that usually has a floppy drive and a DVD or CD drive. Inside the system unit there’s actually another disc drive, which is called the hard drive, and this is where everything “in your computer” is stored.

RAM – Random Access Memory

Almost as common a question as “what is computer hardware?” is “What is RAM?”Abbreviated from “Random Access Memory”, this is the part of the computer that stores whatever it is (files, programs and so on) that you are currently working on.

When you open a file up, the computer basically “moves” it into the RAM memory. The reason the computer does this is that RAM works very fast, much faster

Other Weird Computer Interfaces from CHI 2017

The ACM CHI Conference on Human Factors in Computing Systems is taking place in Denver this week, and just like last year, it’s host to some amazing, incredible, and utterly bizarre technology demos. This year’s theme is “Explore, Innovate, Inspire,” which, as far as we can tell, has no specific meaning and therefore does not constrain the weirdness that CHI is so well known for. We’ve gone through hundreds of 30-second video clips to find the most interesting and craziest stuff, and we can promise you won’t be disappointed. Today, we’re bringing you some interesting ways of interfacing with technology. In addition to videos showing off these breakthroughs, the researchers behind them describe their brainchildren. We’ll have even more videos on virtual and augmented reality and 3D printing later this week

Project Telepathy: Targeted Verbal Communication using 3D Beamforming Speakers and Facial Electromyography

Thanks to the miracle of technology, you can now target individual people and then scream at them without anyone else hearing you, and all it takes it a minimal amount of facial electronification

Anne-Claire Bourland, Peter Gorman, Jess McIntosh, Asier Marzo, University of Bristol, Bristol, United Kingdom

Speech is our innate way of communication. However, it has

Hiring (Software Companies) And Firing (Hardware)

Hiring and Firing announcements show that the year, so far, is a generally good one for software engineers, not so good for their hardware counterparts. And while headline news is incomplete and anecdotal at best, more comprehensive statistics suggest that it represents a trend: software up, hardware down.

Here are layoff and workforce expansion plans made by tech companies so far this year that made the headlines.

In hiring news:

Uber is hiring like crazy in Pittsburgh; according to, it’s looking for 48 engineers, mostly people to work in artificial intelligence and robotics in its advanced technologies group. Recode reported, however, that some of those are replacements for 20 engineers who recently quit as part of a “mini civil war” in the division.

Uber competitor Grab has announced plans to hire 800 R&D engineers over the next two years to staff R&D centers in the United States and Asia, mostly working on machine learning, predictive data analytics, user interfaces, and digital payments.

Didi is hiring self-driving car engineers for its new Mountain View lab, expecting to have “dozens” of researchers working at that facility, focused on AI security and intelligent driving techniques. First, however, the company

Fujitsu Liquid Immersion

Fujitsu on the other hand, is preparing to launch a less exotic solution: a liquid immersion cooling system it says will usher in a “next generation of ultra-dense data centers.”

Though not the first company to come up with the idea, the Japanese computer giant says it’s used its long experience in the field to come up with a design that accommodates both easy maintenance and standard servers. Maintenance is as straightforward to perform as on air-cooled systems, for it does not require gloves, protective clothing or special training, while cables are readily accessible.

Given that liquids are denser than air, Fujitsu says that immersing servers in its new system’s bath of inert fluid greatly improves the cooling process and eliminates the need for server fans. This, in turn, results in a cooling system consuming 40 percent less power compared to that of data centers relying on traditional air-cooling technology. An added bonus is the fanless operation is virtually silent.

“It also reduces the floor space needed by 50 percent,” says Takashi Yamamoto, Vice President, Mechanical & Thermal Engineering Div., Advanced System R&D Unit, Fujitsu. Yamamoto showed off a demonstration system at the company’s annual technology forum

Exascale Will Drive Supercomputing

The United States no longer claimed even the bronze medal. With this week’s release of the latest Top 500 supercomputer ranking, the top three fastest supercomputers in the world are now run by China (with both first and second place finishers) and Switzerland. And while the supercomputer horserace is spectacle enough unto itself, a new report on the supercomputer industry highlights broader trends behind both the latest and the last few years of Top500 rankings.

The report, commissioned last year by the Japanese national science agency Riken, outlines a worldwide race toward exascale computers in which the U.S. sees R&D spending and supercomputer talent pools shrink, Europe jumps into the breach with increased funding, and China pushes hard to become the new global leader, despite a still small user and industry base ready to use the world’s most powerful supercomputers.

Steve Conway, report co-author and senior vice president of research at Hyperion, says the industry trend in high-performance computing is toward laying groundwork for pervasive AI and big data applications like autonomous cars and machine learning. And unlike more specialized supercomputer applications from years past, the workloads of tomorrow’s supercomputers will likely be mainstream and even consumer-facing applications.

“Ten years ago the rationale

The Real Future of Quantum Computing

 Quantum computers based on qubits that can each adopt only two possible options, scientists have now developed a microchip that can generate “qudits” that can each assume 10 or more states, potentially opening up a new way to creating incredibly powerful quantum computers, a new study finds.

Classical computers switch transistors either on or off to symbolize data as ones and zeroes. In contrast, quantum computers use quantum bits, or qubits that, because of the bizarre nature of quantum physics, can be in a state of superposition where they simultaneously act as both 1 and 0.

The superpositions that qubits can adopt let them each help perform two calculations at once. If two qubits are quantum-mechanically linked, or entangled, they can help perform four calculations simultaneously; three qubits, eight calculations; and so on. As a result, a quantum computer with 300 qubits could perform more calculations in an instant than there are atoms in the known universe, solving certain problems much faster than classical computers. However, superpositions are extraordinarily fragile, making it difficult to work with multiple qubits.

Most attempts at building practical quantum computers rely on particles that serve as qubits. However, scientists have long known that they could in principle use qudits with more

The Computer Memory Terminal

The name we give to this experimental information service. It is an attempt to harness the power of the computer in the service of the community. We hope to do this by providing a sort of super bulletin board where people can post notices of all sorts and can find the notices posted by others rapidly.

We are Loving Grace Cybernetics, a group of Berkeley people operating out of Resource One Inc., a non-profit collective located in Project One in S.F. Resource One grew out of the San Francisco Switchboard and has managed to obtain control of a computer (XDS 940) for use in communications.

Pictured above is one of the Community Memory teletype terminals. The first was installed at Leopold’s Records, a student-run record store in Berkeley. The terminal connected by modem to a time-sharing computer in San Francisco, which hosted the electronic bulletin-board system. Users could exchange brief messages about a wide range of topics: apartment listings, music lessons, even where to find a decent bagel. Reading the bulletin board was free, but posting a listing cost a quarter, payable by the coin-op mechanism. The terminals offered many users their first interaction with a computer.

Among the

Complex Biological Computer

Developed a biological computer that functions inside living bacterial cells and tells them what to do, according to a report published today in Nature. Composed of ribonucleic acid, or RNA, the new “ribocomputer” can survive in the bacterium E. coli and respond to a dozen inputs, making it the most complex biological computer to date.

“We’ve developed a way to control how cells behave,” says Alexander Green, an engineer at The Biodesign Institute at Arizona State University, who developed the technology with colleagues at Harvard’s Wyss Institute for Biologically Inspired Engineering. The cells go about their normal business, replicating and sensing what’s going on in their environments, “but they’ve also got this layer of computational machinery that we’ve instructed them to synthesize,” he says.

The biological circuit works just like a digital one: It receives an input and makes a logic-based decision, using AND, OR, and NOT operations. But instead of the inputs and outputs being voltage signals, they are the presence or absence of specific chemicals or proteins.

The process begins with the design of a DNA strand that codes for all the logic the system will need. The researchers insert the synthesized DNA into E. coli bacteria as part of a plasmid—a ring of DNA that can replicate as it floats around in the cell.


The Man Who Brought Style to Supercomputers

A computer that can perform many more calculations per second than the typical computer of its era. The definition is in constant flux: Yesterday’s supercomputer packed the punch of today’s smartphone. From 1969 to 1975, Control Data Corp.’s CDC 7600 was considered the world’s fastest computer, running at 36 megahertz. An iPhone 7, by contrast, runs at 2.33 gigahertz—nearly 100 times as fast as the 7600.

The CDC 7600 was the brainchild of Seymour Cray, who from the 1950s through the 1980s was the undisputed champion among supercomputer designers. Working from a rural laboratory in his hometown of Chippewa Falls, Wisc., Cray had also designed the 7600’s predecessor, the CDC 6600.

Each 7600 had thousands of electronic modules, each built from individual transistors and interconnected with more than 190 kilometers of internal wiring. In the photo above are circular testing points for a few of the first 7600 modules, as seen through the computer’s distinctive blue-glass doors.

Cray viewed the aesthetically pleasing design of the 7600 as a break with the past. As he explains in rare footage documenting the computer’s introduction, earlier computers looked more like “black boxes or gray boxes or white boxes.” For the 7600, Cray chose the blue-glass doors