วันจันทร์ที่ 2 กันยายน พ.ศ. 2556

Electronic Arts's CryEngine 3 [Best Realistic Game Engine]

     If you play Crysis game series, you will know what is the game engine using for made that Crysis.
But now when Crysis 3 arrive, It's use the new CryEngine 3 for made up game. And CryEngine 3 got awards "Game Art" from Game Republic Student Showcase 2012.

   CryEngine 3 Technology 

 

   CryEngine 3 Award "Game Art"


     If you think I want to get CryEngine 3 for made my game, but I don't have any money.
Don't warry CryEngine 3 is a Open SDK and you can get the link from below of this post.
And now you will see how to using CryEnging 3.

Download link for CryEngine 3 : http://mycryengine.com/
How to using CryEngine 3 [Thai] : http://thaicryengine.konlengame.com/


How to Build a Water Cooled PC

      If you are new to liquid cooling, or if you've never purchased Koolance products in particular, you may be wondering what is required to get started. A typical water cooling system consists of four main parts (see also: Liquid Cooling 101):
  1. A Radiator (heat exchanger) with fans to move heat from liquid into air
  2. Water Blocks to transfer heat into liquid
  3. A Pump to move the liquid
  4. A Reservoir for automatically filtering air from the liquid and storing excess coolant
      There are many practical water cooling configurations depending on your application and preferences. You should begin your decision based on which components will be water cooled.      Regardless of whether you're cooling a computer or something else, the expected heat output and desired temperature range of these areas will dictate many of your liquid cooling parts.

Determining Approximate Heat Output
      Hardware is designed with a TDP, or "Thermal Design Power" in mind. This is the maximum amount of heat a cooling system is expected to handle for that component at normal clock speed and voltage. Here is a rough guide:
  • CPU Processor: 60-150W
  • Video Card
    • Single GPU (low-end): 100W
    • Single GPU (mid-range): 150-250W
    • Single GPU (high-end): 200-350W
    • Dual GPU (high-end): 300-450W
  • Motherboard
    • Chipset: 10-30W
    • Voltage Regulators: 5-20W
  • Memory: 2-5W per stick
  • Hard Drive (regular or SSD): 10-30W
      The two primary targets for water cooling in a PC computer are the CPU and video graphics card. These areas produce the highest amount of heat and benefit most from liquid cooling. We can consider these "high heat" sources (a dual-GPU video card should be considered as two high heat sources). The remaining areas on the motherboard, RAM, and hard drives are considered "low heat" sources. Low heat components can be considered in aggregate, but they don't usually contribute enough heat to significantly affect radiator selection.

Choosing a Radiator

radiator 

      Heat exchanger size and airflow are critical to a PC water cooling system's performance-- moreso than liquid flow rate. For this reason, it's recommended to use the largest radiator you can comfortably fit into your workspace, computer chassis, etc. Larger radiators are advantageous because they decrease liquid temperature and can allow for quieter fan speeds.
      What is the minimum radiator size needed if you have space constraints? Our suggested minimum sizes are based on the number of "high heat" devices (CPU or GPU) you will liquid cool:
  • 1 device = 1 fan radiator
  • 2 devices = 2 fan radiator
  • 3 devices = 3 fan radiator
  • 4 devices = 4 fan radiator
  • 5+ devices = larger than 4 fans, or use multiple radiators
      These are only recommendations. The "correct" option is based on your desired temperature and noise range. Some customers find it acceptable to cool 4 video cards with a 3-fan radiator by accepting a somewhat higher temperature range, and/or by running the fans faster. Downsizing too much is something to avoid, though, since it's entirely possible to choose a radiator which is too small to handle a heat load.
      Koolance lists "FPI" (fins per inch) for its heat exchangers, which is the fin density. This can be relevant for users deciding to do one of the following:
  1. Emphasize cooling performance and opt for the largest, highest fin density radiator allowable. Pair it with high CFM/pressure fans. Generally, 120mm fans push more air than 140mm fans.
  2. Emphasize lower noise levels by selecting a lower fin density radiator. Use medium-range fans and/or voltage-throttle them. Generally, 140mm fans are quieter than 120mm fans.
      Low fin density radiators will still improve with more airflow, and high fin density radiators can be quieted by reducing fan speed, so there is a lot of room for tweaking. Either decision should result in significantly lower chip temperatures than air cooling (see recommended radiator sizes above).

Selecting Water Blocks



video card water block

     Koolance has a range of individual water blocks broken down by category. For PC cooling, a convenient Product Selection Tool is also offered. After supplying some basic hardware criteria, this page will generate a list of potential water blocks to use in your future cooling system. Also see our water block help pages under "Information->Product Help" above. If you require assistance, please let us know.

Finding a Pump

 pump 
     Koolance offers several pumps of various specifications. The more cooling components added to a cooling loop, the stronger the pump needed to counter flow restriction. For a typical computer cooling loop with a 3-fan radiator and a few water blocks, any pump offered by Koolance should provide enough flow.
      Flow rate tends to be over-emphasized in PC cooling. For the majority of loops, effective flow rates higher than 1.5-2.0 LPM (0.4-0.5 GPM) won't contribute much, if anything, to thermal performance. A reliable pump is important, as is making sure it's strong enough to keep adequate flow through your selected components. But for users looking to improve thermal performance, increasing radiator size and airflow is almost always more effective.
      Keep in mind that the maximum flow rate listed for pumps is at zero static head pressure, while the maximum static head is at zero flow rate. That means the actual flow rate in a cooling system will usually be quite a bit lower than the pump's maximum specification.

The Reservoir

reservoir
      The primary purpose of a reservoir is to bleed air from the loop and to store extra liquid to reduce maintenance. It won't assist with cooling aside from delaying the time required to reach maximum heat saturation. Reservoirs are also a good opportunity to show off your water cooling system. The size and type of reservoir is based solely on aesthetics and available space. A large, LED-lit reservoir with UV colored coolant mounted against a side window or front drive bay will be highly visible. Hose Size and Fittings (6mm, 10mm, or 13mm?) Tubing is based on allowable space and personal preference. 6mm (1/4in) internal diameter hose is a good option for compact areas like servers and media centers. For computers with more space, 10mm (3/8in) or 13mm (1/2in) ID is recommended. There are few situations where 13mm (1/2in) ID hose outperforms 10mm (3/8in) in temperature, so we encourage this choice based primarily on whichever looks best to you.
      Your fittings will follow the hose size you choose. Be sure that both the ID (internal diameter) and OD (outer diameter) of your fittings match your selected hose size. Hose barbs, unlike compression fittings, will accept different outer diameters by changing the clamp (the ID must still match). Barbs require pliers to install the clamp, while compression fittings are secured by hand.
  
     Water cooling for GPU only.


And this is for CPU water cooling.

Evolution of the Computer

What is a Computer?

     In its most basic form a computer is any device which aids humans in performing various kinds of computations or calculations. In that respect the earliest computer was the abacus, used to perform basic arithmetic operations.
Every computer supports some form of input, processing, and output. This is less obvious on a primitive device such as the abacus where input, output and processing are simply the act of moving the pebbles into new positions, seeing the changed positions, and counting. Regardless, this is what computing is all about, in a nutshell. We input information, the computer processes it according to its basic logic or the program currently running, and outputs the results.
     Modern computers do this electronically, which enables them to perform a vastly greater number of calculations or computations in less time. Despite the fact that we currently use computers to process images, sound, text and other non-numerical forms of data, all of it depends on nothing more than basic numerical calculations. Graphics, sound etc. are merely abstractions of the numbers being crunched within the machine; in digital computers these are the ones and zeros, representing electrical on and off states, and endless combinations of those. In other words every image, every sound, and every word have a corresponding binary code.
While abacus may have technically been the first computer most people today associate the word “computer” with electronic computers which were invented in the last century, and have evolved into modern computers we know of today.

History of The Computer
                   ENIAC
First Generation Computers (1940s – 1950s)
     First electronic computers used vacuum tubes, and they were huge and complex. The first general purpose electronic computer was the ENIAC (Electronic Numerical Integrator And Computer). It was digital, although it didn’t operate with binary code, and was reprogrammable to solve a complete range of computing problems. It was programmed using plugboards and switches, supporting input from an IBM card reader, and output to an IBM card punch. It took up 167 square meters, weighed 27 tons, and consuming 150 kilowatts of power. It used thousands of vacuum tubes, crystal diodes, relays, resistors, and capacitors.
     The first non-general purpose computer was ABC (Atanasoff–Berry Computer), and other similar computers of this era included german Z3, ten British Colossus computers, LEO, Harvard Mark I, and UNIVAC.

History of The Computer
               IBM 1401
Second Generation Computers (1955 – 1960)
     The second generation of computers came about thanks to the invention of the transistor, which then started replacing vacuum tubes in computer design. Transistor computers consumed far less power, produced far less heat, and were much smaller compared to the first generation, albeit still big by today’s standards.
The first transistor computer was created at the University of Manchester in 1953. The most popular of transistor computers was IBM 1401. IBM also created the first disk drive in 1956, the IBM 350 RAMAC.

Third Generation Computers (1960s)

History of The Computer
          IBM System/360

     The invention of the integrated circuits (ICs), also known as microchips, paved the way for computers as we know them today. Making circuits out of single pieces of silicon, which is a semiconductor, allowed them to be much smaller and more practical to produce. This also started the ongoing process of integrating an ever larger number of transistors onto a single microchip. During the sixties microchips started making their way into computers, but the process was gradual, and second generation of computers still held on.
     First appeared minicomputers, first of which were still based on non-microchip transistors, and later versions of which were hybrids, being based on both transistors and microchips, such as IBM’s System/360. They were much smaller, and cheaper than first and second generation of computers, also known as mainframes. Minicomputers can be seen as a bridge between mainframes and microcomputers, which came later as the proliferation of microchips in computers grew.

Fourth Generation Computers (1971 – present)
     First microchips-based central processing units consisted of multiple microchips for different CPU components. The drive for ever greater integration and miniaturization led towards single-chip CPUs, where all of the necessary CPU components were put onto a single microchip, called a microprocessor. The first single-chip CPU, or a microprocessor, was Intel 4004.
The advent of the microprocessor spawned the evolution of the microcomputers, the kind that would eventually become personal computers that we are familiar with today.

First Generation of Microcomputers (1971 – 1976)

History of The Computer
            Altair 8800

     First microcomputers were a weird bunch. They often came in kits, and many were essentially just boxes with lights and switches, usable only to engineers and hobbyists whom could understand binary code. Some, however, did come with a keyboard and/or a monitor, bearing somewhat more resemblance to modern computers.
     It is arguable which of the early microcomputers could be called a first. CTC Datapoint 2200 is one candidate, although it actually didn’t contain a microprocessor (being based on a multi-chip CPU design instead), and wasn’t meant to be a standalone computer, but merely a terminal for the mainframes. The reason some might consider it a first microcomputer is because it could be used as a de-facto standalone computer, it was small enough, and its multi-chip CPU architecture actually became a basis for the x86 architecture later used in IBM PC and its descendants. Plus, it even came with a keyboard and a monitor, an exception in those days.
However, if we are looking for the first microcomputer that came with a proper microprocessor, was meant to be a standalone computer, and didn’t come as a kit then it would be Micral N, which used Intel 8008 microprocessor.
     Popular early microcomputers which did come in kits include MOS Technology KIM-1, Altair 8800, and Apple I. Altair 8800 in particular spawned a large following among the hobbyists, and is considered the spark that started the microcomputer revolution, as these hobbyists went on to found companies centered around personal computing, such as Microsoft, and Apple.

Second Generation Microcomputers (1977 – present)




History of The Computer
       Commodore PET2001

     As microcomputers continued to evolve they became easier to operate, making them accessible to a larger audience. They typically came with a keyboard and a monitor, or could be easily connected to a TV, and they supported visual representation of text and numbers on the screen.
     In other words, lights and switches were replaced by screens and keyboards, and the necessity to understand binary code was diminished as they increasingly came with programs that could be used by issuing more easily understandable commands. Famous early examples of such computers include Commodore PET, Apple II, and in the 80s the IBM PC.
     The nature of the underlying electronic components didn’t change between these computers and modern computers we know of today, but what did change was the number of circuits that could be put onto a single microchip. Intel’s co-founder Gordon Moore predicted the doubling of the number of transistor on a single chip every two years, which became known as “Moore’s Law”, and this trend has roughly held for over 30 years thanks to advancing manufacturing processes and microprocessor designs.
     The consequence was a predictable exponential increase in processing power that could be put into a smaller package, which had a direct effect on the possible form factors as well as applications of modern computers, which is what most of the forthcoming paradigm shifting innovations in computing were about.

Graphical User Interface (GUI)

History of The Computer
      Macintosh 128k

     Possibly the most significant of those shifts was the invention of the graphical user interface, and the mouse as a way of controlling it. Doug Engelbart and his team at the Stanford Research Lab developed the first mouse, and a graphical user interface, demonstrated in 1968. They were just a few years short of the beginning of the personal computer revolution sparked by the Altair 8800 so their idea didn’t take hold.
     Instead it was picked up and improved upon by researchers at the Xerox PARC research center, which in 1973 developed Xerox Alto, the first computer with a mouse-driven GUI. It never became a commercial product, however, as Xerox management wasn’t ready to dive into the computer market and didn’t see the potential of what they had early enough.
     It took Steve Jobs negotiating a stocks deal with Xerox in exchange for a tour of their research center to finally bring the user friendly graphical user interface, as well as the mouse, to the masses. Steve Jobs was shown what Xerox PARC team had developed, and directed Apple to improve upon it. In 1984 Apple introduced the Macintosh, the first mass-market computer with a graphical user interface and a mouse.
     Microsoft later caught on and produced Windows, and the historic competition between the two companies started, resulting in improvements to the graphical user interface to this day.
Meanwhile IBM was dominating the PC market with their IBM PC, and Microsoft was riding on their coat tails by being the one to produce and sell the operating system for the IBM PC known as “DOS” or “Disk Operating System”. Macintosh, with its graphical user interface, was meant to dislodge IBM’s dominance, but Microsoft made this more difficult with their PC-compatible Windows operating system with its own GUI.

Portable Computers

History of The Computer
            Powerbook 150

     As it turned out the idea of a laptop-like portable computer existed even before it was possible to create one, and it was developed at Xerox PARC by Alan Kay whom called it the Dynabook and intended it for children. The first portable computer that was created was the Xerox Notetaker, but only 10 were produced.
     The first laptop that was commercialized was Osborne 1 in 1981, with a small 5″ CRT monitor and a keyboard that sits inside of the lid when closed. It ran CP/M (the OS that Microsoft bought and based DOS on). Later portable computers included Bondwell 2 released in 1985, also running CP/M, which was among the first with a hinge-mounted LCD display. Compaq Portable was the first IBM PC compatible computer, and it ran MS-DOS, but was less portable than Bondwell 2. Other examples of early portable computers included Epson HX-20, GRiD compass, Dulmont Magnum, Kyotronic 85, Commodore SX-64, IBM PC Convertible, Toshiba T1100, T1000, and T1200 etc.
     The first portable computers which resemble modern laptops in features were Apple’s Powerbooks, which first introduced a built-in trackball, and later a trackpad and optional color LCD screens. IBM’s ThinkPad was largely inspired by Powerbook’s design, and the evolution of the two led to laptops and notebook computers as we know them. Powerbooks were eventually replaced by modern MacBook Pro’s.
     Of course, much of the evolution of portable computers was enabled by the evolution of microprocessors, LCD displays, battery technology and so on. This evolution ultimately allowed computers even smaller and more portable than laptops, such as PDAs, tablets, and smartphones.




Network on music instrument.

   If you think the network is only on PC, I will telling you "You are false"
   The network not only on the PC, It can be on telephone, smartphone and it can be on the music instrument. Mostly it on the Live Stage Concert, when you see the concert you can see the Mixing control on down of the stage. That like a Modem of network for everyone on the stage for controlling sound for the match song sound.
   It connecting to the mixing control by the 2 inch - 6.5 inch jack. But it not connecting to mixing control straigth, A jack is connecting with AMP and monitor then it have a 6.5 inch jack from any AMP and monitor to mixing control. And done this network complete. When you playing a music on the stage, staff can mixing your sound to perfect song sound.



Did you see a jack connecting on this Synthesizer. And then it will connecting to the mixing control.

Updating of Google Earth

     Hi! Now we know what is the "Google Earth". And before Google Earth not support at Suphan-Buri, but sometime later Google Earth has been updating and now it support almost all areas of Suphan-Buri. For Example

   
Can you remember here? Yes, it is in fornt of our school Kanchanapisek Wittayalai Suphan Buri.


The turn on fornt of school.


Now, It's a front of our school!!.
 

Suphan Buri Stadium, And the main stage of Suphanburi FC.



วันจันทร์ที่ 12 สิงหาคม พ.ศ. 2556

Electronic Entertainment Expo (E3)

     Electronic Entertainment Expo, commonly known as E3, is an annual trade fair for the computer and video games industry presented by the Entertainment Software Association (ESA). It is used by many video game publishers and accessory manufacturers to show off their upcoming games and game-related merchandise.
Unlike Gamescom and other video game trade fairs that are open to the public, E3 is an exclusive, industry-only event. Persons who apply to attend are required by the event's governing body (Entertainment Software Association) to verify that they have some professional connection to the video game industry.
     E3 is commonly held in late May or early June of each year at the Los Angeles Convention Center (LACC) in Los Angeles. In 2007, the convention was exceptionally held from July 11 to July 13 in Santa Monica, California.
     E3 2013 was held on June 11–13, 2013 at the Los Angeles Convention Center. E3 2014 is slated for June 10–12, 2014, and will once again take place at the Los Angeles Convention Center.

Prior to E3, most game publishers went to other trade shows to display new products, including the Consumer Electronics Show and the European Computer Trade Show.
     
 http://www.digitaltrends.com/wp-content/uploads/2013/06/E3-2013-Preview-header.jpg
            E3 2013 on June 11-13, 2013. The EA Presenting. 

     The first E3 was conceived by IDG's Infotainment World and co-founded by the Interactive Digital Software Association (now the Entertainment Software Association). It coincided with the start of a new generation of consoles, with the release of the Sega Saturn, and the announcements of upcoming releases of the PlayStation, Virtual Boy and Neo-Geo CD. Specifications for the Nintendo Ultra 64 (later renamed Nintendo 64) were released, but there was no hardware shown.
     IDSA originally asked CES for a private meeting space for game developers, but was told that they could not limit access to only invited registrants. Patrick Ferrell, CEO of IDG's Infotainment World, had sent his VP Marketing to the meeting, and hearing the result, the management team at Infotainment World immediately announced E3. Needing to ensure the full backing of the industry, Ferrell then negotiated a partnership between IDG and the IDSA, who then co-produced the show for a number of years
The event ran from May 11 through May 13, 1996 in Los Angeles, California. Keynote speakers included Sega of America, Inc. president and CEO Thomas Kalinske; Sony Electronic Publishing Company president Olaf Olafsson; and Nintendo chairman Howard Lincoln. The first show was one of the largest trade show launches in history, with over 1.2 million feet of show space and over 80,000 attendees.

Electronic Arts (EA)

      Electronic Arts, Inc. (EA) is an American developer, marketer, publisher and distributor of video games. Founded and incorporated on May 28, 1982 by Trip Hawkins, the company was a pioneer of the early home computer games industry and was notable for promoting the designers and programmers responsible for its games. Electronic Arts is the world's third-largest gaming company by revenue after Nintendo and Activision Blizzard.
     Currently, EA develops and publishes games under several labels including EA Sports titles, Madden NFL, FIFA Football/Soccer, NHL, NCAA Football, SSX and NBA Jam. Other EA labels produce established franchises such as Battlefield, Need for Speed, The Sims, Medal of Honor, Command & Conquer, Postal, Mirror's Edge as well as newer franchises such as Dead Space, Mass Effect, Dragon Age, Army of Two and Star Wars: Knights of the Old Republic, produced in partnership with LucasArts. EA also owns and operates major gaming studios, Tiburon in Orlando, EA Canada in Burnaby, BioWare in Edmonton as well as Montreal and DICE in Sweden.

http://www.fighting118th.com/wp-content/uploads/2012/11/mohw-3.jpg
       
        Medal of Honor: Warfighter is one from Electronic Arts.

     Originally, EA was a home computing game publisher. In the late 1980s, the company began developing games in-house and supported consoles by the early 1990s. EA later grew via acquisition of several successful developers. By the early 2000s, EA had become one of the world's largest third-party publishers. On May 4, 2011, EA reported $3.8 billion in revenues for the fiscal year ending March 2011, and on January 13, 2012, EA announced that it had exceeded $1 billion in digital revenue during the previous calendar year. In a note to employees, EA CEO John Riccitiello called this “an incredibly important milestone” for the company. EA began to move toward direct distribution of digital games and services with the acquisition of the popular online gaming site Pogo.com in 2001. In 2009, EA acquired the London-based social gaming startup Playfish, and in June 2011, EA launched Origin, an online service to sell downloadable games directly to consumers. In July 2011, EA announced that it had acquired PopCap Games, the company behind hits such as Plants vs. Zombies and Bejeweled.   

    https://static-www.ecs.popcap.com/www.popcap.com/sites/pvz_franchise_2013/screenshots/gw/PvZ_GW_E3_Screens_04_WM_v3.jpg     "Plants vs. Zombies Garden Warfare" using the new Frostbite 3 game engine. Producer by PopCap Games.

     EA continued its shift toward digital goods in 2012, folding its mobile-focused EA Interactive (EAi) division "into other organizations throughout the company, specifically those divisions led by EA Labels president Frank Gibeau, COO Peter Moore, and CTO Rajat Taneja, and EVP of digital Kristian Segerstrale."

 http://www.thenerdcabinet.com/wp-content/uploads/2013/07/attack-on-titan-shingeki-no-kyojin-attack-on-titan-34606226-1382-900.jpg
     
     You know this right !? Now the best social animation of 2013 
"進撃の巨人" or "Attack on Titan"  is on the producer's consult. 
If producer get the green light to produce this animation to video game, we will see the Attack on Titan game on 2014.