A computer components & hardware forum. HardwareBanter

If this is your first visit, be sure to check out the FAQ by clicking the link above. You may have to register before you can post: click the register link above to proceed. To start viewing messages, select the forum that you want to visit from the selection below.

Go Back   Home » HardwareBanter forum » Video Cards » Nvidia Videocards
Site Map Home Register Authors List Search Today's Posts Mark Forums Read Web Partners

The next Unreal engine...



 
 
Thread Tools Display Modes
  #1  
Old February 26th 04, 05:46 PM
John
external usenet poster
 
Posts: n/a
Default The next Unreal engine...

A good interview with Tim Sweeney on the development of the future
Unreal 3 engine:

http://www.beyond3d.com/interviews/sweeney04/


He says "...we're going to make a game that brings today's GeForce
FX's and Radeon 9700+'s to their knees at 640x480! :-) We are
targetting next-generation consoles and the kinds of PC's that will be
typical on the market in 2006, and today's high end graphics cards are
going to be somewhat low end then, similar to a GeForce4MX or a Radeon
7500 for today's games".

I also like the part where he says he wishes the Intel integrated
graphics chip would just "go away."
  #2  
Old February 26th 04, 06:07 PM
faster_framerates
external usenet poster
 
Posts: n/a
Default

I'm sorry, but what is the benefit of excluding a market of consumers with
average video cards?

How about an engine that runs great and looks beautiful on a large range of
systems? I'm all for progress and a more cinematic look, but Joe Consumer
shouldn't have to upgrade his computer every six months and stay on top of
hardware issues just because he wants to play the latest game release.

This is why people settle for consoles.

- f_f



"John" wrote in message
om...
A good interview with Tim Sweeney on the development of the future
Unreal 3 engine:

http://www.beyond3d.com/interviews/sweeney04/


He says "...we're going to make a game that brings today's GeForce
FX's and Radeon 9700+'s to their knees at 640x480! :-) We are
targetting next-generation consoles and the kinds of PC's that will be
typical on the market in 2006, and today's high end graphics cards are
going to be somewhat low end then, similar to a GeForce4MX or a Radeon
7500 for today's games".

I also like the part where he says he wishes the Intel integrated
graphics chip would just "go away."



  #3  
Old February 26th 04, 06:44 PM
Frank
external usenet poster
 
Posts: n/a
Default

id love to see the gameplay evolving as much as they plan
to do with the grafix. im not going to buy me one of those cards
just to play another bunnyhopping fragfest. jaw drops quickly
with grafix and fx, but the need to play the game again and
again has as much or more to do with a decent gameplay.

cu


frank


  #4  
Old February 26th 04, 07:04 PM
external usenet poster
 
Posts: n/a
Default

In alt.comp.periphs.videocards.ati faster_framerates wrote:
I'm sorry, but what is the benefit of excluding a market of consumers
with average video cards?


They are targeting average video cards, it's just that they're targeting
the average cards of 2006. It will be closer to 2006 when the engine is
finished, and targeting even high-end cards of 2003 in a project that
starts in 2004 is just a waste of time and money. A 24-month upgrade
cycle is not completely unreasonable for videogames.

This is why people settle for consoles.


I take issue with the "settle", but that's another argument for another
time.

-a
  #5  
Old February 26th 04, 08:11 PM
John Lewis
external usenet poster
 
Posts: n/a
Default

On Thu, 26 Feb 2004 19:44:59 +0100, "Frank" wrote:

id love to see the gameplay evolving as much as they plan
to do with the grafix. im not going to buy me one of those cards
just to play another bunnyhopping fragfest. jaw drops quickly
with grafix and fx, but the need to play the game again and
again has as much or more to do with a decent gameplay.


And in 3D open-space games such as H&D2, Far Cry etc,
a whole lot to do with AI as well. Replaying a level with
scripted AI, and knowing exactly where the enemy will
appear and how they will react gives zero re-play value.
So a vote for "intelligent AI", such as that in Far Cry,
H&D2 also seem to have elements of intelligent AI.

And, of course there is no substitute for clever and
interesting level-design with unexpected game-play
and plot twists having multiple endings, such as
Deus Ex 1.

John Lewis

cu


frank



  #6  
Old February 26th 04, 11:41 PM
Kevin C.
external usenet poster
 
Posts: n/a
Default


wrote in message
...
In alt.comp.periphs.videocards.ati faster_framerates

wrote:
I'm sorry, but what is the benefit of excluding a market of consumers
with average video cards?


They are targeting average video cards, it's just that they're targeting
the average cards of 2006.


That's somewhat contradictory with his statement that he wishes the Intel
video chips would go away. Whether you choose to believe it or not, most
people do not own high end GPUs today, nor will they tomorrow. Even above
the casual gamer, there are many folks who are still running GF2-era
devices. In 2006 I imagine that just as many people will still be running
the GF4s and Radeons that are in their computers today, the same cards that
Mr. Sweeney has targeted to exclude.


  #7  
Old February 26th 04, 11:53 PM
Bora Ugurlu
external usenet poster
 
Posts: n/a
Default

Am Thu, 26 Feb 2004 23:41:36 GMT, "Kevin C."

That's somewhat contradictory with his statement that he wishes the Intel
video chips would go away. Whether you choose to believe it or not, most
people do not own high end GPUs today, nor will they tomorrow. Even above
the casual gamer, there are many folks who are still running GF2-era
devices. In 2006 I imagine that just as many people will still be running
the GF4s and Radeons that are in their computers today, the same cards that
Mr. Sweeney has targeted to exclude.



He meant to say 'at maximum detail level'.. Then the high end cards
would come to a crawl. Not many people play with details maxed out. So
they get in, say, Ut2k4 decent framerates with a Ti4200 (which I
have). If I turn on all the details with 4xAA and 8x Anisotropy then
it's a slide show. That's what he meant.
  #8  
Old February 27th 04, 01:08 AM
K
external usenet poster
 
Posts: n/a
Default


"Kevin C." wrote in message
om...

That's somewhat contradictory with his statement that he wishes the Intel
video chips would go away. Whether you choose to believe it or not, most
people do not own high end GPUs today, nor will they tomorrow. Even above
the casual gamer, there are many folks who are still running GF2-era
devices. In 2006 I imagine that just as many people will still be running
the GF4s and Radeons that are in their computers today, the same cards

that
Mr. Sweeney has targeted to exclude.


Well sucks to be them. It's about time software started pushing the limits
of hardware again. There was a time when people were very happy to get
30fps from Quake 2. Now all you see is people concerned that they are only

getting 90 fps in UT2003, etc. If in 2006 people still choose to hold on to
their GF4s and Radeons they are going to be left out on new titles, and they
only have themselves to blame. You cannot expect the software developers to
stand still for the benefit of those who are unwilling to upgrade.

There has only been two occasions when I've installed a gfx card and said
'wow' to myself. The first was playing Unreal and Q2 on a Voodoo 2, the
other was after I got a GF3 and seen all the Q3 engined games in high-res
with all the candy. All the cards since then have only done what the GF3
did, just faster. In other words there has been little in the way of
innovation. What has been long overdue in the graphics industry is a next
'wow' card.


K


  #9  
Old February 27th 04, 01:57 AM
rms
external usenet poster
 
Posts: n/a
Default

And in 3D open-space games such as H&D2, Far Cry etc,
a whole lot to do with AI as well. Replaying a level with
scripted AI, and knowing exactly where the enemy will
appear and how they will react gives zero re-play value.


Playing Vietcong I'm continually astounded in the variety of AI
placement and behavior that this game offers. For instance I'm now trying
to complete one of the quickfights (Arroyo). It's quite difficult, and I've
restarted the level literally dozens of times, and each time the initial AI
placement and behavior is slightly different. Very impressive.

rms


  #10  
Old February 27th 04, 06:59 AM
Andrew
external usenet poster
 
Posts: n/a
Default

On Fri, 27 Feb 2004 01:08:25 -0000, "K" wrote:

There has only been two occasions when I've installed a gfx card and said
'wow' to myself. The first was playing Unreal and Q2 on a Voodoo 2, the
other was after I got a GF3 and seen all the Q3 engined games in high-res
with all the candy. All the cards since then have only done what the GF3
did, just faster. In other words there has been little in the way of
innovation. What has been long overdue in the graphics industry is a next
'wow' card.


Far Cry on a 9700 Pro graphics card gave me a "wow". Even seeing the
rain on water in Morrowind in a GF4 was a "wow" moment for me. There
has been a lot of innovation in hardware and software since the GF3.
--
Andrew. To email unscramble & remove spamtrap.
Help make Usenet a better place: English is read downwards,
please don't top post. Trim messages to quote only relevent text.
Check groups.google.com before asking a question.
 




Thread Tools
Display Modes

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

vB code is On
Smilies are On
[IMG] code is On
HTML code is Off
Forum Jump

Similar Threads
Thread Thread Starter Forum Replies Last Post
"Sysmark" and "Unreal" definitions ml wahl Homebuilt PC's 1 November 17th 04 12:16 AM
The next Unreal engine... John Ati Videocards 41 February 29th 04 07:05 AM
Unreal 1 and a softmodded 9500-9700 radeon [email protected] writeme.com Ati Videocards 2 August 28th 03 04:38 PM
Asus 9280(128MB) low 3D performance with 3DMark2001 and Unreal Tournament Hugo Sondermeijer Nvidia Videocards 3 July 27th 03 04:46 PM
What's P200 Mhx that Unreal Tournament asks for? remeb General 1 July 3rd 03 06:52 AM


All times are GMT +1. The time now is 06:07 PM.


Powered by vBulletin® Version 3.6.4
Copyright ©2000 - 2022, Jelsoft Enterprises Ltd.
Copyright 2004-2022 HardwareBanter.
The comments are property of their posters.